% Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 3831 100 3831 0 0 10261 0 --:--:-- --:--:-- --:--:-- 10270 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 9315 100 9315 0 0 23176 0 --:--:-- --:--:-- --:--:-- 23171 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 2607 100 2607 0 0 7800 0 --:--:-- --:--:-- --:--:-- 7805 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 1953 100 1953 0 0 5877 0 --:--:-- --:--:-- --:--:-- 5882 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 350 100 350 0 0 998 0 --:--:-- --:--:-- --:--:-- 997 Installing kuttl Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/kubectl-kuttl https://github.com/kudobuilder/kuttl/releases/download/v0.15.0/kubectl-kuttl_0.15.0_linux_x86_64 KUBECONFIG file is: /tmp/kubeconfig-4053090600 for suite in elasticsearch examples generate upgrade sidecar streaming ui miscellaneous; do \ make run-e2e-tests-$suite ; \ done make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh elasticsearch false true + '[' 3 -ne 3 ']' + test_suite_name=elasticsearch + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/elasticsearch.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-elasticsearch make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true \ KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ ./tests/e2e/elasticsearch/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-25-163410 True False 7m28s Cluster version is 4.18.0-0.nightly-2025-01-25-163410' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-25-163410 True False 7m28s Cluster version is 4.18.0-0.nightly-2025-01-25-163410' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/elasticsearch/render.sh ++ export SUITE_DIR=./tests/e2e/elasticsearch ++ SUITE_DIR=./tests/e2e/elasticsearch ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/elasticsearch ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + start_test es-from-aio-to-production + '[' 1 -ne 1 ']' + test_name=es-from-aio-to-production + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-from-aio-to-production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-from-aio-to-production\e[0m' Rendering files for test es-from-aio-to-production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-from-aio-to-production + cd es-from-aio-to-production + jaeger_name=my-jaeger + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 03 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=03 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./03-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch.redundancyPolicy="ZeroRedundancy"' ./03-install.yaml + render_smoke_test my-jaeger true 04 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=04 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./04-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test es-increasing-replicas + '[' 1 -ne 1 ']' + test_name=es-increasing-replicas + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-increasing-replicas' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-increasing-replicas\e[0m' Rendering files for test es-increasing-replicas + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-from-aio-to-production + '[' es-from-aio-to-production '!=' _build ']' + cd .. + mkdir -p es-increasing-replicas + cd es-increasing-replicas + jaeger_name=simple-prod + '[' true = true ']' + jaeger_deployment_mode=production_autoprovisioned + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.replicas=2 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.query.replicas=2 ./02-install.yaml + cp ./01-assert.yaml ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.replicas=2 ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .status.readyReplicas=2 ./02-assert.yaml + render_smoke_test simple-prod true 03 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=03 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./03-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + cp ./02-install.yaml ./04-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.elasticsearch.nodeCount=2 ./04-install.yaml + /tmp/jaeger-tests/bin/gomplate -f ./openshift-check-es-nodes.yaml.template -o ./05-check-es-nodes.yaml + '[' true = true ']' + skip_test es-index-cleaner-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-increasing-replicas + '[' es-increasing-replicas '!=' _build ']' + cd .. + rm -rf es-index-cleaner-upstream + warning 'es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_index_cleaner -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-index-cleaner-autoprov + '[' 1 -ne 1 ']' + test_name=es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-index-cleaner-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-index-cleaner-autoprov\e[0m' Rendering files for test es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-index-cleaner-autoprov + cd es-index-cleaner-autoprov + jaeger_name=test-es-index-cleaner-with-prefix + cronjob_name=test-es-index-cleaner-with-prefix-es-index-cleaner + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + cp ../../es-index-cleaner-upstream/04-assert.yaml ../../es-index-cleaner-upstream/README.md . + render_install_jaeger test-es-index-cleaner-with-prefix production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options.es.index-prefix=""' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.enabled=false ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.numberOfDays=0 ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.esIndexCleaner.schedule="*/1 * * * *"' ./01-install.yaml + render_report_spans test-es-index-cleaner-with-prefix true 5 00 true 02 + '[' 6 -ne 6 ']' + jaeger=test-es-index-cleaner-with-prefix + is_secured=true + number_of_spans=5 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + export JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=5 + DAYS=5 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + sed 's~enabled: false~enabled: true~gi' ./01-install.yaml + CRONJOB_NAME=test-es-index-cleaner-with-prefix-es-index-cleaner + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./04-wait-es-index-cleaner.yaml + /tmp/jaeger-tests/bin/gomplate -f ./01-install.yaml -o ./05-install.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 00 06 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=00 + test_step=06 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=test-es-index-cleaner-with-prefix-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=test-es-index-cleaner-with-prefix-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./06-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./06-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.16"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.16"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.16"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.16 ++ version_ge 5.8.16 5.4 +++ echo 5.8.16 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.16 == 5.8.16 + '[' -n '' ']' + skip_test es-index-cleaner-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-index-cleaner-autoprov + '[' es-index-cleaner-autoprov '!=' _build ']' + cd .. + rm -rf es-index-cleaner-managed + warning 'es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + start_test es-multiinstance + '[' 1 -ne 1 ']' + test_name=es-multiinstance + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-multiinstance' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-multiinstance\e[0m' Rendering files for test es-multiinstance + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-multiinstance + cd es-multiinstance + jaeger_name=instance-1 + render_install_jaeger instance-1 production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=instance-1 + JAEGER_NAME=instance-1 + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f ./03-create-second-instance.yaml.template -o 03-create-second-instance.yaml + '[' true = true ']' + skip_test es-rollover-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-rollover-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-multiinstance + '[' es-multiinstance '!=' _build ']' + cd .. + rm -rf es-rollover-upstream + warning 'es-rollover-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_rollover -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-rollover-autoprov + '[' 1 -ne 1 ']' + test_name=es-rollover-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-rollover-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-rollover-autoprov\e[0m' Rendering files for test es-rollover-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-rollover-autoprov + cd es-rollover-autoprov + cp ../../es-rollover-upstream/05-assert.yaml ../../es-rollover-upstream/05-install.yaml ../../es-rollover-upstream/README.md . + jaeger_name=my-jaeger + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_report_spans my-jaeger true 2 00 true 02 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 00 03 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=00 + test_step=03 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./03-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./03-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 01 04 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=01 + test_step=04 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=01 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./04-check-indices.yaml + JOB_NUMBER=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./04-assert.yaml + render_report_spans my-jaeger true 2 02 true 06 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=02 + ensure_reported_spans=true + test_step=06 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=02 + JOB_NUMBER=02 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./06-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./06-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 02 07 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=02 + test_step=07 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=02 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./07-check-indices.yaml + JOB_NUMBER=02 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./07-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' 03 08 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + job_number=03 + test_step=08 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=03 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./08-check-indices.yaml + JOB_NUMBER=03 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./08-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' 04 09 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + job_number=04 + test_step=09 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=04 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./09-check-indices.yaml + JOB_NUMBER=04 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./09-assert.yaml + render_report_spans my-jaeger true 2 03 true 10 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=03 + ensure_reported_spans=true + test_step=10 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=03 + JOB_NUMBER=03 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./10-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./10-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + CRONJOB_NAME=my-jaeger-es-rollover + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./11-wait-rollover.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-000002'\'',' 05 11 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-000002'\'',' + job_number=05 + test_step=11 + escape_command ''\''--name'\'', '\''jaeger-span-000002'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-000002'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-000002'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-000002'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=05 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./11-check-indices.yaml + JOB_NUMBER=05 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./11-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' 06 12 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + job_number=06 + test_step=12 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=06 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./12-check-indices.yaml + JOB_NUMBER=06 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./12-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.16"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.16"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.16"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.16 ++ version_ge 5.8.16 5.4 +++ echo 5.8.16 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.16 == 5.8.16 + '[' -n '' ']' + skip_test es-rollover-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-rollover-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-rollover-autoprov + '[' es-rollover-autoprov '!=' _build ']' + cd .. + rm -rf es-rollover-managed + warning 'es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + skip_test es-spark-dependencies 'This test is not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=es-spark-dependencies + message='This test is not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + rm -rf es-spark-dependencies + warning 'es-spark-dependencies: This test is not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-spark-dependencies: This test is not supported in OpenShift\e[0m' WAR: es-spark-dependencies: This test is not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running elasticsearch E2E tests' Running elasticsearch E2E tests + cd tests/e2e/elasticsearch/_build + set +e + KUBECONFIG=/tmp/kubeconfig-4053090600 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/es-from-aio-to-production === PAUSE kuttl/harness/es-from-aio-to-production === RUN kuttl/harness/es-increasing-replicas === PAUSE kuttl/harness/es-increasing-replicas === RUN kuttl/harness/es-index-cleaner-autoprov === PAUSE kuttl/harness/es-index-cleaner-autoprov === RUN kuttl/harness/es-multiinstance === PAUSE kuttl/harness/es-multiinstance === RUN kuttl/harness/es-rollover-autoprov === PAUSE kuttl/harness/es-rollover-autoprov === RUN kuttl/harness/es-simple-prod === PAUSE kuttl/harness/es-simple-prod === CONT kuttl/harness/artifacts logger.go:42: 07:18:21 | artifacts | Creating namespace: kuttl-test-handy-oriole logger.go:42: 07:18:21 | artifacts | artifacts events from ns kuttl-test-handy-oriole: logger.go:42: 07:18:21 | artifacts | Deleting namespace: kuttl-test-handy-oriole === CONT kuttl/harness/es-multiinstance logger.go:42: 07:18:27 | es-multiinstance | Ignoring 03-create-second-instance.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:18:27 | es-multiinstance | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:18:27 | es-multiinstance | Creating namespace: kuttl-test-mint-monster logger.go:42: 07:18:27 | es-multiinstance/0-clear-namespace | starting test step 0-clear-namespace logger.go:42: 07:18:27 | es-multiinstance/0-clear-namespace | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --ignore-not-found=true] logger.go:42: 07:18:27 | es-multiinstance/0-clear-namespace | test step completed 0-clear-namespace logger.go:42: 07:18:27 | es-multiinstance/1-install | starting test step 1-install logger.go:42: 07:18:27 | es-multiinstance/1-install | Jaeger:kuttl-test-mint-monster/instance-1 created logger.go:42: 07:19:26 | es-multiinstance/1-install | test step completed 1-install logger.go:42: 07:19:26 | es-multiinstance/2-create-namespace | starting test step 2-create-namespace logger.go:42: 07:19:26 | es-multiinstance/2-create-namespace | running command: [sh -c kubectl create namespace jaeger-e2e-multiinstance-test] logger.go:42: 07:19:26 | es-multiinstance/2-create-namespace | namespace/jaeger-e2e-multiinstance-test created logger.go:42: 07:19:26 | es-multiinstance/2-create-namespace | test step completed 2-create-namespace logger.go:42: 07:19:26 | es-multiinstance/3-create-second-instance | starting test step 3-create-second-instance logger.go:42: 07:19:26 | es-multiinstance/3-create-second-instance | running command: [sh -c kubectl apply -f ./01-install.yaml -n jaeger-e2e-multiinstance-test] logger.go:42: 07:19:26 | es-multiinstance/3-create-second-instance | jaeger.jaegertracing.io/instance-1 created logger.go:42: 07:19:26 | es-multiinstance/3-create-second-instance | running command: [sh -c /tmp/jaeger-tests/bin/kubectl-kuttl assert ./01-assert.yaml -n jaeger-e2e-multiinstance-test --timeout 1000] logger.go:42: 07:20:20 | es-multiinstance/3-create-second-instance | assert is valid logger.go:42: 07:20:20 | es-multiinstance/3-create-second-instance | test step completed 3-create-second-instance logger.go:42: 07:20:20 | es-multiinstance/4-check-secrets | starting test step 4-check-secrets logger.go:42: 07:20:20 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n $NAMESPACE > secret1] logger.go:42: 07:20:20 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n jaeger-e2e-multiinstance-test > secret2] logger.go:42: 07:20:20 | es-multiinstance/4-check-secrets | running command: [sh -c cmp --silent secret1 secret2 || exit 0] logger.go:42: 07:20:20 | es-multiinstance/4-check-secrets | test step completed 4-check-secrets logger.go:42: 07:20:20 | es-multiinstance/5-delete | starting test step 5-delete logger.go:42: 07:20:20 | es-multiinstance/5-delete | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --wait=false] logger.go:42: 07:20:20 | es-multiinstance/5-delete | namespace "jaeger-e2e-multiinstance-test" deleted logger.go:42: 07:20:20 | es-multiinstance/5-delete | test step completed 5-delete logger.go:42: 07:20:20 | es-multiinstance | es-multiinstance events from ns kuttl-test-mint-monster: logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:18:35 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestmintmonsterinstance1-1-875f857c7 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestmintmonsterinstance1-1-875f857cdndc9 replicaset-controller logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:18:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmintmonsterinstance1-1-875f857cdndc9 Binding Scheduled Successfully assigned kuttl-test-mint-monster/elasticsearch-cdm-kuttltestmintmonsterinstance1-1-875f857cdndc9 to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:18:35 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestmintmonsterinstance1-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestmintmonsterinstance1-1-875f857c7 to 1 deployment-controller logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:18:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmintmonsterinstance1-1-875f857cdndc9 AddedInterface Add eth0 [10.129.2.19/23] from ovn-kubernetes multus logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:18:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmintmonsterinstance1-1-875f857cdndc9.spec.containers{elasticsearch} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:18:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmintmonsterinstance1-1-875f857cdndc9.spec.containers{elasticsearch} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" in 7.593s (7.593s including waiting). Image size: 539381180 bytes. kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:18:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmintmonsterinstance1-1-875f857cdndc9.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:18:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmintmonsterinstance1-1-875f857cdndc9.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:18:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmintmonsterinstance1-1-875f857cdndc9.spec.containers{proxy} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:18:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmintmonsterinstance1-1-875f857cdndc9.spec.containers{proxy} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" in 3.899s (3.899s including waiting). Image size: 287327885 bytes. kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:18:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmintmonsterinstance1-1-875f857cdndc9.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:18:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmintmonsterinstance1-1-875f857cdndc9.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:18:56 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmintmonsterinstance1-1-875f857cdndc9.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:01 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmintmonsterinstance1-1-875f857cdndc9.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:12 +0000 UTC Normal Pod instance-1-collector-557984c787-n7wjt Binding Scheduled Successfully assigned kuttl-test-mint-monster/instance-1-collector-557984c787-n7wjt to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:12 +0000 UTC Normal ReplicaSet.apps instance-1-collector-557984c787 SuccessfulCreate Created pod: instance-1-collector-557984c787-n7wjt replicaset-controller logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:12 +0000 UTC Normal Deployment.apps instance-1-collector ScalingReplicaSet Scaled up replica set instance-1-collector-557984c787 to 1 deployment-controller logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:12 +0000 UTC Normal Pod instance-1-query-b5bd795c5-jmb7w Binding Scheduled Successfully assigned kuttl-test-mint-monster/instance-1-query-b5bd795c5-jmb7w to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:12 +0000 UTC Normal ReplicaSet.apps instance-1-query-b5bd795c5 SuccessfulCreate Created pod: instance-1-query-b5bd795c5-jmb7w replicaset-controller logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:12 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-b5bd795c5 to 1 deployment-controller logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:13 +0000 UTC Normal Pod instance-1-collector-557984c787-n7wjt AddedInterface Add eth0 [10.131.0.29/23] from ovn-kubernetes multus logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:13 +0000 UTC Normal Pod instance-1-collector-557984c787-n7wjt.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:13 +0000 UTC Normal Pod instance-1-query-b5bd795c5-jmb7w AddedInterface Add eth0 [10.128.2.16/23] from ovn-kubernetes multus logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:13 +0000 UTC Normal Pod instance-1-query-b5bd795c5-jmb7w.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:17 +0000 UTC Normal Pod instance-1-collector-557984c787-n7wjt.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" in 4.054s (4.054s including waiting). Image size: 142489237 bytes. kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:17 +0000 UTC Normal Pod instance-1-collector-557984c787-n7wjt.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:17 +0000 UTC Normal Pod instance-1-collector-557984c787-n7wjt.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:18 +0000 UTC Normal Pod instance-1-query-b5bd795c5-jmb7w.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" in 5.45s (5.45s including waiting). Image size: 192936692 bytes. kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:18 +0000 UTC Normal Pod instance-1-query-b5bd795c5-jmb7w.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:18 +0000 UTC Normal Pod instance-1-query-b5bd795c5-jmb7w.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:18 +0000 UTC Normal Pod instance-1-query-b5bd795c5-jmb7w.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:21 +0000 UTC Normal Pod instance-1-query-b5bd795c5-jmb7w.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" in 2.458s (2.458s including waiting). Image size: 339954870 bytes. kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:21 +0000 UTC Normal Pod instance-1-query-b5bd795c5-jmb7w.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:21 +0000 UTC Normal Pod instance-1-query-b5bd795c5-jmb7w.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:21 +0000 UTC Normal Pod instance-1-query-b5bd795c5-jmb7w.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:23 +0000 UTC Normal Pod instance-1-query-b5bd795c5-jmb7w.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" in 2.471s (2.471s including waiting). Image size: 115697219 bytes. kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:23 +0000 UTC Normal Pod instance-1-query-b5bd795c5-jmb7w.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:24 +0000 UTC Normal Pod instance-1-query-b5bd795c5-jmb7w.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod instance-1-collector-557984c787-n7wjt horizontal-pod-autoscaler logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:19:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:20:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod instance-1-collector-557984c787-n7wjt horizontal-pod-autoscaler logger.go:42: 07:20:20 | es-multiinstance | 2025-01-27 07:20:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod instance-1-collector-557984c787-n7wjt horizontal-pod-autoscaler logger.go:42: 07:20:20 | es-multiinstance | Deleting namespace: kuttl-test-mint-monster === CONT kuttl/harness/es-simple-prod logger.go:42: 07:20:27 | es-simple-prod | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:20:27 | es-simple-prod | Creating namespace: kuttl-test-loyal-rooster logger.go:42: 07:20:27 | es-simple-prod | es-simple-prod events from ns kuttl-test-loyal-rooster: logger.go:42: 07:20:27 | es-simple-prod | Deleting namespace: kuttl-test-loyal-rooster === CONT kuttl/harness/es-rollover-autoprov logger.go:42: 07:20:33 | es-rollover-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:20:33 | es-rollover-autoprov | Creating namespace: kuttl-test-brave-beetle logger.go:42: 07:20:33 | es-rollover-autoprov/1-install | starting test step 1-install logger.go:42: 07:20:33 | es-rollover-autoprov/1-install | Jaeger:kuttl-test-brave-beetle/my-jaeger created logger.go:42: 07:21:20 | es-rollover-autoprov/1-install | test step completed 1-install logger.go:42: 07:21:20 | es-rollover-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:21:20 | es-rollover-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:21:21 | es-rollover-autoprov/2-report-spans | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:21:29 | es-rollover-autoprov/2-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:21:29 | es-rollover-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:21:29 | es-rollover-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:21:57 | es-rollover-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 07:21:57 | es-rollover-autoprov/3-check-indices | starting test step 3-check-indices logger.go:42: 07:21:57 | es-rollover-autoprov/3-check-indices | Job:kuttl-test-brave-beetle/00-check-indices created logger.go:42: 07:22:02 | es-rollover-autoprov/3-check-indices | test step completed 3-check-indices logger.go:42: 07:22:02 | es-rollover-autoprov/4-check-indices | starting test step 4-check-indices logger.go:42: 07:22:02 | es-rollover-autoprov/4-check-indices | Job:kuttl-test-brave-beetle/01-check-indices created logger.go:42: 07:22:08 | es-rollover-autoprov/4-check-indices | test step completed 4-check-indices logger.go:42: 07:22:08 | es-rollover-autoprov/5-install | starting test step 5-install logger.go:42: 07:22:08 | es-rollover-autoprov/5-install | Jaeger:kuttl-test-brave-beetle/my-jaeger updated logger.go:42: 07:22:15 | es-rollover-autoprov/5-install | test step completed 5-install logger.go:42: 07:22:15 | es-rollover-autoprov/6-report-spans | starting test step 6-report-spans logger.go:42: 07:22:15 | es-rollover-autoprov/6-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:22:22 | es-rollover-autoprov/6-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=02 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-02-job.yaml] logger.go:42: 07:22:23 | es-rollover-autoprov/6-report-spans | running command: [sh -c kubectl apply -f report-span-02-job.yaml -n $NAMESPACE] logger.go:42: 07:22:23 | es-rollover-autoprov/6-report-spans | job.batch/02-report-span created logger.go:42: 07:22:47 | es-rollover-autoprov/6-report-spans | test step completed 6-report-spans logger.go:42: 07:22:47 | es-rollover-autoprov/7-check-indices | starting test step 7-check-indices logger.go:42: 07:22:47 | es-rollover-autoprov/7-check-indices | Job:kuttl-test-brave-beetle/02-check-indices created logger.go:42: 07:22:51 | es-rollover-autoprov/7-check-indices | test step completed 7-check-indices logger.go:42: 07:22:51 | es-rollover-autoprov/8-check-indices | starting test step 8-check-indices logger.go:42: 07:22:51 | es-rollover-autoprov/8-check-indices | Job:kuttl-test-brave-beetle/03-check-indices created logger.go:42: 07:22:56 | es-rollover-autoprov/8-check-indices | test step completed 8-check-indices logger.go:42: 07:22:56 | es-rollover-autoprov/9-check-indices | starting test step 9-check-indices logger.go:42: 07:22:56 | es-rollover-autoprov/9-check-indices | Job:kuttl-test-brave-beetle/04-check-indices created logger.go:42: 07:23:00 | es-rollover-autoprov/9-check-indices | test step completed 9-check-indices logger.go:42: 07:23:00 | es-rollover-autoprov/10-report-spans | starting test step 10-report-spans logger.go:42: 07:23:00 | es-rollover-autoprov/10-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:23:09 | es-rollover-autoprov/10-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=03 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-03-job.yaml] logger.go:42: 07:23:09 | es-rollover-autoprov/10-report-spans | running command: [sh -c kubectl apply -f report-span-03-job.yaml -n $NAMESPACE] logger.go:42: 07:23:10 | es-rollover-autoprov/10-report-spans | job.batch/03-report-span created logger.go:42: 07:23:35 | es-rollover-autoprov/10-report-spans | test step completed 10-report-spans logger.go:42: 07:23:35 | es-rollover-autoprov/11-check-indices | starting test step 11-check-indices logger.go:42: 07:23:35 | es-rollover-autoprov/11-check-indices | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob my-jaeger-es-rollover --namespace $NAMESPACE] logger.go:42: 07:23:44 | es-rollover-autoprov/11-check-indices | time="2025-01-27T07:23:44Z" level=debug msg="Checking if the my-jaeger-es-rollover CronJob exists" logger.go:42: 07:23:44 | es-rollover-autoprov/11-check-indices | time="2025-01-27T07:23:44Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 07:23:44 | es-rollover-autoprov/11-check-indices | time="2025-01-27T07:23:44Z" level=info msg="Cronjob my-jaeger-es-rollover found successfully" logger.go:42: 07:23:44 | es-rollover-autoprov/11-check-indices | time="2025-01-27T07:23:44Z" level=debug msg="Waiting for the next scheduled job from my-jaeger-es-rollover cronjob" logger.go:42: 07:23:44 | es-rollover-autoprov/11-check-indices | time="2025-01-27T07:23:44Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:23:54 | es-rollover-autoprov/11-check-indices | time="2025-01-27T07:23:54Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:24:04 | es-rollover-autoprov/11-check-indices | time="2025-01-27T07:24:04Z" level=info msg="Job of owner my-jaeger-es-rollover succeeded after my-jaeger-es-rollover 20.016713752s" logger.go:42: 07:24:04 | es-rollover-autoprov/11-check-indices | Job:kuttl-test-brave-beetle/05-check-indices created logger.go:42: 07:24:09 | es-rollover-autoprov/11-check-indices | test step completed 11-check-indices logger.go:42: 07:24:09 | es-rollover-autoprov/12-check-indices | starting test step 12-check-indices logger.go:42: 07:24:09 | es-rollover-autoprov/12-check-indices | Job:kuttl-test-brave-beetle/06-check-indices created logger.go:42: 07:24:13 | es-rollover-autoprov/12-check-indices | test step completed 12-check-indices logger.go:42: 07:24:13 | es-rollover-autoprov | es-rollover-autoprov events from ns kuttl-test-brave-beetle: logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:20:40 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestbravebeetlemyjaeger-1-7b6f99575c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestbravebeetlemyjaeger-1-7b6f99575l79sb replicaset-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:20:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbravebeetlemyjaeger-1-7b6f99575l79sb Binding Scheduled Successfully assigned kuttl-test-brave-beetle/elasticsearch-cdm-kuttltestbravebeetlemyjaeger-1-7b6f99575l79sb to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:20:40 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestbravebeetlemyjaeger-1-7b6f99575l79sb FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:20:40 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestbravebeetlemyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestbravebeetlemyjaeger-1-7b6f99575c to 1 deployment-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:20:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbravebeetlemyjaeger-1-7b6f99575l79sb AddedInterface Add eth0 [10.129.2.20/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:20:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbravebeetlemyjaeger-1-7b6f99575l79sb.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:20:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbravebeetlemyjaeger-1-7b6f99575l79sb.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:20:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbravebeetlemyjaeger-1-7b6f99575l79sb.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:20:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbravebeetlemyjaeger-1-7b6f99575l79sb.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:20:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbravebeetlemyjaeger-1-7b6f99575l79sb.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:20:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbravebeetlemyjaeger-1-7b6f99575l79sb.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:20:56 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestbravebeetlemyjaeger-1-7b6f99575l79sb.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:07 +0000 UTC Normal Pod my-jaeger-collector-f7d5d7d54-v9jpp Binding Scheduled Successfully assigned kuttl-test-brave-beetle/my-jaeger-collector-f7d5d7d54-v9jpp to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:07 +0000 UTC Warning Pod my-jaeger-collector-f7d5d7d54-v9jpp FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:07 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-f7d5d7d54 SuccessfulCreate Created pod: my-jaeger-collector-f7d5d7d54-v9jpp replicaset-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:07 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-f7d5d7d54 to 1 deployment-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:07 +0000 UTC Normal Pod my-jaeger-query-755776f976-xxhb8 Binding Scheduled Successfully assigned kuttl-test-brave-beetle/my-jaeger-query-755776f976-xxhb8 to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:07 +0000 UTC Warning Pod my-jaeger-query-755776f976-xxhb8 FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-oauth-proxy-tls" : secret "my-jaeger-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:07 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-755776f976 SuccessfulCreate Created pod: my-jaeger-query-755776f976-xxhb8 replicaset-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:07 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-755776f976 to 1 deployment-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:08 +0000 UTC Normal Pod my-jaeger-collector-f7d5d7d54-v9jpp AddedInterface Add eth0 [10.128.2.19/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:08 +0000 UTC Normal Pod my-jaeger-collector-f7d5d7d54-v9jpp.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:08 +0000 UTC Normal Pod my-jaeger-collector-f7d5d7d54-v9jpp.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:08 +0000 UTC Normal Pod my-jaeger-collector-f7d5d7d54-v9jpp.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:08 +0000 UTC Normal Pod my-jaeger-query-755776f976-xxhb8 AddedInterface Add eth0 [10.131.0.31/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:08 +0000 UTC Normal Pod my-jaeger-query-755776f976-xxhb8.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:11 +0000 UTC Normal Pod my-jaeger-query-755776f976-xxhb8.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" in 2.748s (2.748s including waiting). Image size: 192936692 bytes. kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:11 +0000 UTC Normal Pod my-jaeger-query-755776f976-xxhb8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:11 +0000 UTC Normal Pod my-jaeger-query-755776f976-xxhb8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:11 +0000 UTC Normal Pod my-jaeger-query-755776f976-xxhb8.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:15 +0000 UTC Normal Pod my-jaeger-query-755776f976-xxhb8.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" in 4.221s (4.221s including waiting). Image size: 339954870 bytes. kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:16 +0000 UTC Normal Pod my-jaeger-query-755776f976-xxhb8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:16 +0000 UTC Normal Pod my-jaeger-query-755776f976-xxhb8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:16 +0000 UTC Normal Pod my-jaeger-query-755776f976-xxhb8.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:18 +0000 UTC Normal Pod my-jaeger-query-755776f976-xxhb8.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" in 2.393s (2.393s including waiting). Image size: 115697219 bytes. kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:18 +0000 UTC Normal Pod my-jaeger-query-755776f976-xxhb8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:18 +0000 UTC Normal Pod my-jaeger-query-755776f976-xxhb8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:22 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:22 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:22 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:24 +0000 UTC Normal Pod my-jaeger-query-755776f976-xxhb8.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:24 +0000 UTC Normal Pod my-jaeger-query-755776f976-xxhb8.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:24 +0000 UTC Normal Pod my-jaeger-query-755776f976-xxhb8.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:24 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-755776f976 SuccessfulDelete Deleted pod: my-jaeger-query-755776f976-xxhb8 replicaset-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:24 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-755776f976 to 0 from 1 deployment-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:25 +0000 UTC Normal Pod my-jaeger-query-575686d5fd-cczkf Binding Scheduled Successfully assigned kuttl-test-brave-beetle/my-jaeger-query-575686d5fd-cczkf to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:25 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-575686d5fd SuccessfulCreate Created pod: my-jaeger-query-575686d5fd-cczkf replicaset-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:25 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-575686d5fd to 1 deployment-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:26 +0000 UTC Normal Pod my-jaeger-query-575686d5fd-cczkf AddedInterface Add eth0 [10.131.0.32/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:26 +0000 UTC Normal Pod my-jaeger-query-575686d5fd-cczkf.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:26 +0000 UTC Normal Pod my-jaeger-query-575686d5fd-cczkf.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:26 +0000 UTC Normal Pod my-jaeger-query-575686d5fd-cczkf.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:26 +0000 UTC Normal Pod my-jaeger-query-575686d5fd-cczkf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:26 +0000 UTC Normal Pod my-jaeger-query-575686d5fd-cczkf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:26 +0000 UTC Normal Pod my-jaeger-query-575686d5fd-cczkf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:26 +0000 UTC Normal Pod my-jaeger-query-575686d5fd-cczkf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:26 +0000 UTC Normal Pod my-jaeger-query-575686d5fd-cczkf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:26 +0000 UTC Normal Pod my-jaeger-query-575686d5fd-cczkf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:29 +0000 UTC Normal Pod 00-report-span-2s5l5 Binding Scheduled Successfully assigned kuttl-test-brave-beetle/00-report-span-2s5l5 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:29 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-2s5l5 job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:30 +0000 UTC Normal Pod 00-report-span-2s5l5 AddedInterface Add eth0 [10.128.2.20/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:30 +0000 UTC Normal Pod 00-report-span-2s5l5.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:33 +0000 UTC Normal Pod 00-report-span-2s5l5.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 3.082s (3.082s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:33 +0000 UTC Normal Pod 00-report-span-2s5l5.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:33 +0000 UTC Normal Pod 00-report-span-2s5l5.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:52 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-f7d5d7d54-v9jpp horizontal-pod-autoscaler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:52 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-f7d5d7d54-v9jpp horizontal-pod-autoscaler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:52 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-f7d5d7d54-v9jpp horizontal-pod-autoscaler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:56 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:57 +0000 UTC Normal Pod 00-check-indices-t5ll6 Binding Scheduled Successfully assigned kuttl-test-brave-beetle/00-check-indices-t5ll6 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:57 +0000 UTC Normal Pod 00-check-indices-t5ll6 AddedInterface Add eth0 [10.128.2.21/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:57 +0000 UTC Normal Pod 00-check-indices-t5ll6.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:57 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-t5ll6 job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:58 +0000 UTC Normal Pod 00-check-indices-t5ll6.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 820ms (820ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:58 +0000 UTC Normal Pod 00-check-indices-t5ll6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:21:58 +0000 UTC Normal Pod 00-check-indices-t5ll6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:01 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:02 +0000 UTC Normal Pod 01-check-indices-47jqm Binding Scheduled Successfully assigned kuttl-test-brave-beetle/01-check-indices-47jqm to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:02 +0000 UTC Normal Pod 01-check-indices-47jqm AddedInterface Add eth0 [10.128.2.22/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:02 +0000 UTC Normal Pod 01-check-indices-47jqm.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:02 +0000 UTC Normal Job.batch 01-check-indices SuccessfulCreate Created pod: 01-check-indices-47jqm job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:04 +0000 UTC Normal Pod 01-check-indices-47jqm.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.06s (2.061s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:04 +0000 UTC Normal Pod 01-check-indices-47jqm.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:04 +0000 UTC Normal Pod 01-check-indices-47jqm.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:07 +0000 UTC Normal Job.batch 01-check-indices Completed Job completed job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:09 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-xwhw9 Binding Scheduled Successfully assigned kuttl-test-brave-beetle/my-jaeger-es-rollover-create-mapping-xwhw9 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:09 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping SuccessfulCreate Created pod: my-jaeger-es-rollover-create-mapping-xwhw9 job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:10 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-xwhw9 AddedInterface Add eth0 [10.128.2.23/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:10 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-xwhw9.spec.containers{my-jaeger-es-rollover-create-mapping} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:12 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-xwhw9.spec.containers{my-jaeger-es-rollover-create-mapping} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" in 2.128s (2.128s including waiting). Image size: 111742913 bytes. kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:12 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-xwhw9.spec.containers{my-jaeger-es-rollover-create-mapping} Created Created container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:12 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-xwhw9.spec.containers{my-jaeger-es-rollover-create-mapping} Started Started container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:15 +0000 UTC Normal Pod my-jaeger-collector-f7d5d7d54-v9jpp.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:15 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-f7d5d7d54 SuccessfulDelete Deleted pod: my-jaeger-collector-f7d5d7d54-v9jpp replicaset-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:15 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-f7d5d7d54 to 0 from 1 deployment-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:15 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping Completed Job completed job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:15 +0000 UTC Normal Pod my-jaeger-query-575686d5fd-cczkf.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:15 +0000 UTC Normal Pod my-jaeger-query-575686d5fd-cczkf.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:15 +0000 UTC Normal Pod my-jaeger-query-575686d5fd-cczkf.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:15 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-575686d5fd SuccessfulDelete Deleted pod: my-jaeger-query-575686d5fd-cczkf replicaset-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:15 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-575686d5fd to 0 from 1 deployment-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:16 +0000 UTC Normal Pod my-jaeger-collector-7b4c4976ff-qvsc8 Binding Scheduled Successfully assigned kuttl-test-brave-beetle/my-jaeger-collector-7b4c4976ff-qvsc8 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:16 +0000 UTC Normal Pod my-jaeger-collector-7b4c4976ff-qvsc8 AddedInterface Add eth0 [10.128.2.24/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:16 +0000 UTC Normal Pod my-jaeger-collector-7b4c4976ff-qvsc8.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:16 +0000 UTC Normal Pod my-jaeger-collector-7b4c4976ff-qvsc8.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:16 +0000 UTC Normal Pod my-jaeger-collector-7b4c4976ff-qvsc8.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:16 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7b4c4976ff SuccessfulCreate Created pod: my-jaeger-collector-7b4c4976ff-qvsc8 replicaset-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:16 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7b4c4976ff to 1 deployment-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:16 +0000 UTC Normal Pod my-jaeger-query-c6fb67999-lc7h2 Binding Scheduled Successfully assigned kuttl-test-brave-beetle/my-jaeger-query-c6fb67999-lc7h2 to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:16 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-c6fb67999 SuccessfulCreate Created pod: my-jaeger-query-c6fb67999-lc7h2 replicaset-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:16 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-c6fb67999 to 1 deployment-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:17 +0000 UTC Normal Pod my-jaeger-query-c6fb67999-lc7h2 AddedInterface Add eth0 [10.131.0.33/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:17 +0000 UTC Normal Pod my-jaeger-query-c6fb67999-lc7h2.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:17 +0000 UTC Normal Pod my-jaeger-query-c6fb67999-lc7h2.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:17 +0000 UTC Normal Pod my-jaeger-query-c6fb67999-lc7h2.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:17 +0000 UTC Normal Pod my-jaeger-query-c6fb67999-lc7h2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:17 +0000 UTC Normal Pod my-jaeger-query-c6fb67999-lc7h2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:17 +0000 UTC Normal Pod my-jaeger-query-c6fb67999-lc7h2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:17 +0000 UTC Normal Pod my-jaeger-query-c6fb67999-lc7h2.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:17 +0000 UTC Normal Pod my-jaeger-query-c6fb67999-lc7h2.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:17 +0000 UTC Normal Pod my-jaeger-query-c6fb67999-lc7h2.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:23 +0000 UTC Normal Pod 02-report-span-r7h6g Binding Scheduled Successfully assigned kuttl-test-brave-beetle/02-report-span-r7h6g to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:23 +0000 UTC Normal Pod 02-report-span-r7h6g AddedInterface Add eth0 [10.128.2.25/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:23 +0000 UTC Normal Pod 02-report-span-r7h6g.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:23 +0000 UTC Normal Job.batch 02-report-span SuccessfulCreate Created pod: 02-report-span-r7h6g job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:24 +0000 UTC Normal Pod 02-report-span-r7h6g.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 874ms (874ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:24 +0000 UTC Normal Pod 02-report-span-r7h6g.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:24 +0000 UTC Normal Pod 02-report-span-r7h6g.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:47 +0000 UTC Normal Pod 02-check-indices-l489j Binding Scheduled Successfully assigned kuttl-test-brave-beetle/02-check-indices-l489j to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:47 +0000 UTC Normal Pod 02-check-indices-l489j AddedInterface Add eth0 [10.128.2.26/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:47 +0000 UTC Normal Job.batch 02-check-indices SuccessfulCreate Created pod: 02-check-indices-l489j job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:47 +0000 UTC Normal Job.batch 02-report-span Completed Job completed job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:48 +0000 UTC Normal Pod 02-check-indices-l489j.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:48 +0000 UTC Normal Pod 02-check-indices-l489j.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 776ms (776ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:48 +0000 UTC Normal Pod 02-check-indices-l489j.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:48 +0000 UTC Normal Pod 02-check-indices-l489j.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:51 +0000 UTC Normal Job.batch 02-check-indices Completed Job completed job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:51 +0000 UTC Normal Pod 03-check-indices-tgdtx Binding Scheduled Successfully assigned kuttl-test-brave-beetle/03-check-indices-tgdtx to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:51 +0000 UTC Normal Job.batch 03-check-indices SuccessfulCreate Created pod: 03-check-indices-tgdtx job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:52 +0000 UTC Normal Pod 03-check-indices-tgdtx AddedInterface Add eth0 [10.128.2.27/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:52 +0000 UTC Normal Pod 03-check-indices-tgdtx.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:53 +0000 UTC Normal Pod 03-check-indices-tgdtx.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.373s (1.373s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:53 +0000 UTC Normal Pod 03-check-indices-tgdtx.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:53 +0000 UTC Normal Pod 03-check-indices-tgdtx.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:56 +0000 UTC Normal Job.batch 03-check-indices Completed Job completed job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:56 +0000 UTC Normal Pod 04-check-indices-vpbj9 Binding Scheduled Successfully assigned kuttl-test-brave-beetle/04-check-indices-vpbj9 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:56 +0000 UTC Normal Job.batch 04-check-indices SuccessfulCreate Created pod: 04-check-indices-vpbj9 job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:57 +0000 UTC Normal Pod 04-check-indices-vpbj9 AddedInterface Add eth0 [10.128.2.28/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:57 +0000 UTC Normal Pod 04-check-indices-vpbj9.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:57 +0000 UTC Normal Pod 04-check-indices-vpbj9.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 727ms (727ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:57 +0000 UTC Normal Pod 04-check-indices-vpbj9.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:22:57 +0000 UTC Normal Pod 04-check-indices-vpbj9.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:00 +0000 UTC Normal Job.batch 04-check-indices Completed Job completed job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28966043-fjlzn Binding Scheduled Successfully assigned kuttl-test-brave-beetle/my-jaeger-es-lookback-28966043-fjlzn to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28966043-fjlzn AddedInterface Add eth0 [10.128.2.29/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28966043-fjlzn.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" already present on machine kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28966043-fjlzn.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28966043-fjlzn.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28966043 SuccessfulCreate Created pod: my-jaeger-es-lookback-28966043-fjlzn job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28966043 cronjob-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28966043-8l94t Binding Scheduled Successfully assigned kuttl-test-brave-beetle/my-jaeger-es-rollover-28966043-8l94t to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28966043-8l94t AddedInterface Add eth0 [10.128.2.30/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28966043-8l94t.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" already present on machine kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28966043-8l94t.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28966043-8l94t.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28966043 SuccessfulCreate Created pod: my-jaeger-es-rollover-28966043-8l94t job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28966043 cronjob-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28966043 Completed Job completed job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28966043, condition: Complete cronjob-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28966043 Completed Job completed job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28966043, condition: Complete cronjob-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:08 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-7b4c4976ff-qvsc8 horizontal-pod-autoscaler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:08 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-7b4c4976ff-qvsc8 horizontal-pod-autoscaler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:08 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-7b4c4976ff-qvsc8 horizontal-pod-autoscaler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:10 +0000 UTC Normal Pod 03-report-span-h7895 Binding Scheduled Successfully assigned kuttl-test-brave-beetle/03-report-span-h7895 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:10 +0000 UTC Normal Pod 03-report-span-h7895 AddedInterface Add eth0 [10.128.2.31/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:10 +0000 UTC Normal Pod 03-report-span-h7895.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:10 +0000 UTC Normal Job.batch 03-report-span SuccessfulCreate Created pod: 03-report-span-h7895 job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:12 +0000 UTC Normal Pod 03-report-span-h7895.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.403s (1.403s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:12 +0000 UTC Normal Pod 03-report-span-h7895.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:12 +0000 UTC Normal Pod 03-report-span-h7895.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:23:34 +0000 UTC Normal Job.batch 03-report-span Completed Job completed job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28966044-dbm5p Binding Scheduled Successfully assigned kuttl-test-brave-beetle/my-jaeger-es-lookback-28966044-dbm5p to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28966044-dbm5p AddedInterface Add eth0 [10.128.2.32/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28966044-dbm5p.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" already present on machine kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28966044-dbm5p.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28966044-dbm5p.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28966044 SuccessfulCreate Created pod: my-jaeger-es-lookback-28966044-dbm5p job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28966044 cronjob-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28966044-dhfqp Binding Scheduled Successfully assigned kuttl-test-brave-beetle/my-jaeger-es-rollover-28966044-dhfqp to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28966044-dhfqp AddedInterface Add eth0 [10.128.2.33/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28966044-dhfqp.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" already present on machine kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28966044-dhfqp.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28966044-dhfqp.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28966044 SuccessfulCreate Created pod: my-jaeger-es-rollover-28966044-dhfqp job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28966044 cronjob-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28966044 Completed Job completed job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28966044, condition: Complete cronjob-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28966044 Completed Job completed job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28966044, condition: Complete cronjob-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:04 +0000 UTC Normal Pod 05-check-indices-6hfxw Binding Scheduled Successfully assigned kuttl-test-brave-beetle/05-check-indices-6hfxw to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:04 +0000 UTC Normal Job.batch 05-check-indices SuccessfulCreate Created pod: 05-check-indices-6hfxw job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:05 +0000 UTC Normal Pod 05-check-indices-6hfxw AddedInterface Add eth0 [10.128.2.34/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:05 +0000 UTC Normal Pod 05-check-indices-6hfxw.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:06 +0000 UTC Normal Pod 05-check-indices-6hfxw.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.178s (1.178s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:06 +0000 UTC Normal Pod 05-check-indices-6hfxw.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:06 +0000 UTC Normal Pod 05-check-indices-6hfxw.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:09 +0000 UTC Normal Job.batch 05-check-indices Completed Job completed job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:09 +0000 UTC Normal Pod 06-check-indices-kw88m Binding Scheduled Successfully assigned kuttl-test-brave-beetle/06-check-indices-kw88m to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:09 +0000 UTC Normal Job.batch 06-check-indices SuccessfulCreate Created pod: 06-check-indices-kw88m job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:10 +0000 UTC Normal Pod 06-check-indices-kw88m AddedInterface Add eth0 [10.128.2.35/23] from ovn-kubernetes multus logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:10 +0000 UTC Normal Pod 06-check-indices-kw88m.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:11 +0000 UTC Normal Pod 06-check-indices-kw88m.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 905ms (905ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:11 +0000 UTC Normal Pod 06-check-indices-kw88m.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:11 +0000 UTC Normal Pod 06-check-indices-kw88m.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:24:13 | es-rollover-autoprov | 2025-01-27 07:24:13 +0000 UTC Normal Job.batch 06-check-indices Completed Job completed job-controller logger.go:42: 07:24:13 | es-rollover-autoprov | Deleting namespace: kuttl-test-brave-beetle === CONT kuttl/harness/es-increasing-replicas logger.go:42: 07:24:20 | es-increasing-replicas | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:24:20 | es-increasing-replicas | Ignoring check-es-nodes.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:24:20 | es-increasing-replicas | Ignoring openshift-check-es-nodes.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:24:20 | es-increasing-replicas | Creating namespace: kuttl-test-pro-panda logger.go:42: 07:24:21 | es-increasing-replicas/1-install | starting test step 1-install logger.go:42: 07:24:21 | es-increasing-replicas/1-install | Jaeger:kuttl-test-pro-panda/simple-prod created logger.go:42: 07:24:57 | es-increasing-replicas/1-install | test step completed 1-install logger.go:42: 07:24:57 | es-increasing-replicas/2-install | starting test step 2-install logger.go:42: 07:24:57 | es-increasing-replicas/2-install | Jaeger:kuttl-test-pro-panda/simple-prod updated logger.go:42: 07:25:04 | es-increasing-replicas/2-install | test step completed 2-install logger.go:42: 07:25:04 | es-increasing-replicas/3-smoke-test | starting test step 3-smoke-test logger.go:42: 07:25:04 | es-increasing-replicas/3-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:25:05 | es-increasing-replicas/3-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:25:11 | es-increasing-replicas/3-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:25:12 | es-increasing-replicas/3-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:25:12 | es-increasing-replicas/3-smoke-test | job.batch/report-span created logger.go:42: 07:25:12 | es-increasing-replicas/3-smoke-test | job.batch/check-span created logger.go:42: 07:25:26 | es-increasing-replicas/3-smoke-test | test step completed 3-smoke-test logger.go:42: 07:25:26 | es-increasing-replicas/4-install | starting test step 4-install logger.go:42: 07:25:26 | es-increasing-replicas/4-install | Jaeger:kuttl-test-pro-panda/simple-prod updated logger.go:42: 07:25:26 | es-increasing-replicas/4-install | test step completed 4-install logger.go:42: 07:25:26 | es-increasing-replicas/5-check-es-nodes | starting test step 5-check-es-nodes logger.go:42: 07:25:26 | es-increasing-replicas/5-check-es-nodes | running command: [sh -c ./check-es-nodes.sh $NAMESPACE] logger.go:42: 07:25:26 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:25:26 | es-increasing-replicas/5-check-es-nodes | false logger.go:42: 07:25:26 | es-increasing-replicas/5-check-es-nodes | Error: no matches found logger.go:42: 07:25:31 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:25:31 | es-increasing-replicas/5-check-es-nodes | true logger.go:42: 07:25:31 | es-increasing-replicas/5-check-es-nodes | test step completed 5-check-es-nodes logger.go:42: 07:25:31 | es-increasing-replicas | es-increasing-replicas events from ns kuttl-test-pro-panda: logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpropandasimpleprod-1-d9957b549-hxqlg Binding Scheduled Successfully assigned kuttl-test-pro-panda/elasticsearch-cdm-kuttltestpropandasimpleprod-1-d9957b549-hxqlg to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:26 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestpropandasimpleprod-1-d9957b549 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestpropandasimpleprod-1-d9957b549-hxqlg replicaset-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:26 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestpropandasimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestpropandasimpleprod-1-d9957b549 to 1 deployment-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpropandasimpleprod-1-d9957b549-hxqlg AddedInterface Add eth0 [10.129.2.21/23] from ovn-kubernetes multus logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpropandasimpleprod-1-d9957b549-hxqlg.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpropandasimpleprod-1-d9957b549-hxqlg.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpropandasimpleprod-1-d9957b549-hxqlg.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpropandasimpleprod-1-d9957b549-hxqlg.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpropandasimpleprod-1-d9957b549-hxqlg.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpropandasimpleprod-1-d9957b549-hxqlg.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:42 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpropandasimpleprod-1-d9957b549-hxqlg.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:53 +0000 UTC Normal Pod simple-prod-collector-64c4745956-r6fn9 Binding Scheduled Successfully assigned kuttl-test-pro-panda/simple-prod-collector-64c4745956-r6fn9 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:53 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-64c4745956 SuccessfulCreate Created pod: simple-prod-collector-64c4745956-r6fn9 replicaset-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:53 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-64c4745956 to 1 deployment-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:53 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-pzzmg Binding Scheduled Successfully assigned kuttl-test-pro-panda/simple-prod-query-6b48d76c68-pzzmg to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:53 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6b48d76c68 SuccessfulCreate Created pod: simple-prod-query-6b48d76c68-pzzmg replicaset-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:53 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6b48d76c68 to 1 deployment-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:54 +0000 UTC Normal Pod simple-prod-collector-64c4745956-r6fn9 AddedInterface Add eth0 [10.128.2.36/23] from ovn-kubernetes multus logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:54 +0000 UTC Normal Pod simple-prod-collector-64c4745956-r6fn9.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:54 +0000 UTC Normal Pod simple-prod-collector-64c4745956-r6fn9.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:54 +0000 UTC Normal Pod simple-prod-collector-64c4745956-r6fn9.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:54 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-pzzmg AddedInterface Add eth0 [10.131.0.34/23] from ovn-kubernetes multus logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:54 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-pzzmg.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:54 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-pzzmg.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:54 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-pzzmg.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:54 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-pzzmg.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:54 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-pzzmg.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:54 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-pzzmg.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:54 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-pzzmg.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:54 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-pzzmg.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:24:54 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-pzzmg.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:00 +0000 UTC Normal Pod simple-prod-collector-64c4745956-wvwtb Binding Scheduled Successfully assigned kuttl-test-pro-panda/simple-prod-collector-64c4745956-wvwtb to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:00 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-64c4745956 SuccessfulCreate Created pod: simple-prod-collector-64c4745956-wvwtb replicaset-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:00 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-64c4745956 to 2 from 1 deployment-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:00 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-gfxf7 Binding Scheduled Successfully assigned kuttl-test-pro-panda/simple-prod-query-6b48d76c68-gfxf7 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:00 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6b48d76c68 SuccessfulCreate Created pod: simple-prod-query-6b48d76c68-gfxf7 replicaset-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:00 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6b48d76c68 to 2 from 1 deployment-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:01 +0000 UTC Normal Pod simple-prod-collector-64c4745956-wvwtb AddedInterface Add eth0 [10.131.0.35/23] from ovn-kubernetes multus logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:01 +0000 UTC Normal Pod simple-prod-collector-64c4745956-wvwtb.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:01 +0000 UTC Normal Pod simple-prod-collector-64c4745956-wvwtb.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:01 +0000 UTC Normal Pod simple-prod-collector-64c4745956-wvwtb.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:01 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-gfxf7 AddedInterface Add eth0 [10.128.2.37/23] from ovn-kubernetes multus logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:01 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-gfxf7.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:01 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-gfxf7.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:01 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-gfxf7.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:01 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-gfxf7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:01 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-gfxf7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:01 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-gfxf7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:01 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-gfxf7.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:01 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-gfxf7.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:01 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-gfxf7.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:07 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-gfxf7.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:07 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-gfxf7.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:07 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-gfxf7.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:07 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-pzzmg.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:07 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-pzzmg.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:07 +0000 UTC Normal Pod simple-prod-query-6b48d76c68-pzzmg.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:07 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6b48d76c68 SuccessfulDelete Deleted pod: simple-prod-query-6b48d76c68-pzzmg replicaset-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:07 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6b48d76c68 SuccessfulDelete Deleted pod: simple-prod-query-6b48d76c68-gfxf7 replicaset-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:07 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-6b48d76c68 to 0 from 2 deployment-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:08 +0000 UTC Normal Pod simple-prod-query-74d4595654-qg9bg Binding Scheduled Successfully assigned kuttl-test-pro-panda/simple-prod-query-74d4595654-qg9bg to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:08 +0000 UTC Normal Pod simple-prod-query-74d4595654-sm7r7 Binding Scheduled Successfully assigned kuttl-test-pro-panda/simple-prod-query-74d4595654-sm7r7 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:08 +0000 UTC Normal ReplicaSet.apps simple-prod-query-74d4595654 SuccessfulCreate Created pod: simple-prod-query-74d4595654-qg9bg replicaset-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:08 +0000 UTC Normal ReplicaSet.apps simple-prod-query-74d4595654 SuccessfulCreate Created pod: simple-prod-query-74d4595654-sm7r7 replicaset-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:08 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-74d4595654 to 2 deployment-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-qg9bg AddedInterface Add eth0 [10.131.0.36/23] from ovn-kubernetes multus logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-qg9bg.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-qg9bg.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-qg9bg.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-qg9bg.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-qg9bg.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-qg9bg.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-qg9bg.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-qg9bg.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-qg9bg.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-sm7r7 AddedInterface Add eth0 [10.128.2.38/23] from ovn-kubernetes multus logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-sm7r7.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-sm7r7.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-sm7r7.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-sm7r7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-sm7r7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-sm7r7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-sm7r7.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-sm7r7.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:09 +0000 UTC Normal Pod simple-prod-query-74d4595654-sm7r7.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:12 +0000 UTC Normal Pod check-span-9jctq Binding Scheduled Successfully assigned kuttl-test-pro-panda/check-span-9jctq to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:12 +0000 UTC Normal Pod check-span-9jctq AddedInterface Add eth0 [10.131.0.37/23] from ovn-kubernetes multus logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:12 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-9jctq job-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:12 +0000 UTC Normal Pod report-span-m7rjq Binding Scheduled Successfully assigned kuttl-test-pro-panda/report-span-m7rjq to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:12 +0000 UTC Normal Pod report-span-m7rjq AddedInterface Add eth0 [10.128.2.39/23] from ovn-kubernetes multus logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:12 +0000 UTC Normal Pod report-span-m7rjq.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:12 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-m7rjq job-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:13 +0000 UTC Normal Pod check-span-9jctq.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:13 +0000 UTC Normal Pod report-span-m7rjq.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 796ms (796ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:13 +0000 UTC Normal Pod report-span-m7rjq.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:13 +0000 UTC Normal Pod report-span-m7rjq.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:14 +0000 UTC Normal Pod check-span-9jctq.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.647s (1.647s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:14 +0000 UTC Normal Pod check-span-9jctq.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:14 +0000 UTC Normal Pod check-span-9jctq.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:25 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:27 +0000 UTC Normal Pod simple-prod-collector-64c4745956-r6fn9.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:27 +0000 UTC Normal Pod simple-prod-collector-64c4745956-wvwtb.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:27 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-64c4745956 SuccessfulDelete Deleted pod: simple-prod-collector-64c4745956-wvwtb replicaset-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:27 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-64c4745956 SuccessfulDelete Deleted pod: simple-prod-collector-64c4745956-r6fn9 replicaset-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:27 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled down replica set simple-prod-collector-64c4745956 to 0 from 2 deployment-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:27 +0000 UTC Normal Pod simple-prod-query-74d4595654-qg9bg.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:27 +0000 UTC Normal Pod simple-prod-query-74d4595654-qg9bg.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:27 +0000 UTC Normal Pod simple-prod-query-74d4595654-qg9bg.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:27 +0000 UTC Normal Pod simple-prod-query-74d4595654-sm7r7.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:27 +0000 UTC Normal Pod simple-prod-query-74d4595654-sm7r7.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:27 +0000 UTC Normal Pod simple-prod-query-74d4595654-sm7r7.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:27 +0000 UTC Normal ReplicaSet.apps simple-prod-query-74d4595654 SuccessfulDelete Deleted pod: simple-prod-query-74d4595654-qg9bg replicaset-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:27 +0000 UTC Normal ReplicaSet.apps simple-prod-query-74d4595654 SuccessfulDelete Deleted pod: simple-prod-query-74d4595654-sm7r7 replicaset-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:27 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-74d4595654 to 0 from 2 deployment-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestpropandasimpleprod-2-86bf8df85f SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestpropandasimpleprod-2-86bf8df85fcssxw replicaset-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpropandasimpleprod-2-86bf8df85fcssxw Binding Scheduled Successfully assigned kuttl-test-pro-panda/elasticsearch-cdm-kuttltestpropandasimpleprod-2-86bf8df85fcssxw to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpropandasimpleprod-2-86bf8df85fcssxw AddedInterface Add eth0 [10.131.0.38/23] from ovn-kubernetes multus logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpropandasimpleprod-2-86bf8df85fcssxw.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpropandasimpleprod-2-86bf8df85fcssxw.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpropandasimpleprod-2-86bf8df85fcssxw.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpropandasimpleprod-2-86bf8df85fcssxw.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpropandasimpleprod-2-86bf8df85fcssxw.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpropandasimpleprod-2-86bf8df85fcssxw.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestpropandasimpleprod-2 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestpropandasimpleprod-2-86bf8df85f to 1 deployment-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal Pod simple-prod-collector-544bfdb879-srl95 Binding Scheduled Successfully assigned kuttl-test-pro-panda/simple-prod-collector-544bfdb879-srl95 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal Pod simple-prod-collector-544bfdb879-v8ldc Binding Scheduled Successfully assigned kuttl-test-pro-panda/simple-prod-collector-544bfdb879-v8ldc to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-544bfdb879 SuccessfulCreate Created pod: simple-prod-collector-544bfdb879-srl95 replicaset-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-544bfdb879 SuccessfulCreate Created pod: simple-prod-collector-544bfdb879-v8ldc replicaset-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-544bfdb879 to 2 deployment-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal Pod simple-prod-query-7c97565969-bvj99 Binding Scheduled Successfully assigned kuttl-test-pro-panda/simple-prod-query-7c97565969-bvj99 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal Pod simple-prod-query-7c97565969-c94w8 Binding Scheduled Successfully assigned kuttl-test-pro-panda/simple-prod-query-7c97565969-c94w8 to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal ReplicaSet.apps simple-prod-query-7c97565969 SuccessfulCreate Created pod: simple-prod-query-7c97565969-bvj99 replicaset-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal ReplicaSet.apps simple-prod-query-7c97565969 SuccessfulCreate Created pod: simple-prod-query-7c97565969-c94w8 replicaset-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:28 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-7c97565969 to 2 deployment-controller logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-collector-544bfdb879-srl95 AddedInterface Add eth0 [10.128.2.40/23] from ovn-kubernetes multus logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-collector-544bfdb879-srl95.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-collector-544bfdb879-srl95.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-collector-544bfdb879-srl95.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-collector-544bfdb879-v8ldc AddedInterface Add eth0 [10.131.0.39/23] from ovn-kubernetes multus logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-collector-544bfdb879-v8ldc.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-collector-544bfdb879-v8ldc.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-collector-544bfdb879-v8ldc.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-bvj99 AddedInterface Add eth0 [10.128.2.41/23] from ovn-kubernetes multus logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-bvj99.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-bvj99.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-bvj99.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-bvj99.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-bvj99.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-bvj99.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-bvj99.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-bvj99.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-bvj99.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-c94w8 AddedInterface Add eth0 [10.131.0.40/23] from ovn-kubernetes multus logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-c94w8.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-c94w8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-c94w8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-c94w8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-c94w8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-c94w8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-c94w8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-c94w8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:25:31 | es-increasing-replicas | 2025-01-27 07:25:29 +0000 UTC Normal Pod simple-prod-query-7c97565969-c94w8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:25:31 | es-increasing-replicas | Deleting namespace: kuttl-test-pro-panda === CONT kuttl/harness/es-index-cleaner-autoprov logger.go:42: 07:26:06 | es-index-cleaner-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:26:06 | es-index-cleaner-autoprov | Creating namespace: kuttl-test-cool-guinea logger.go:42: 07:26:06 | es-index-cleaner-autoprov/1-install | starting test step 1-install logger.go:42: 07:26:06 | es-index-cleaner-autoprov/1-install | Jaeger:kuttl-test-cool-guinea/test-es-index-cleaner-with-prefix created logger.go:42: 07:26:42 | es-index-cleaner-autoprov/1-install | test step completed 1-install logger.go:42: 07:26:42 | es-index-cleaner-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:26:42 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE test-es-index-cleaner-with-prefix /dev/null] logger.go:42: 07:26:44 | es-index-cleaner-autoprov/2-report-spans | Warning: resource jaegers/test-es-index-cleaner-with-prefix is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:26:51 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c DAYS=5 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:26:52 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:26:52 | es-index-cleaner-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:36:52 | es-index-cleaner-autoprov/2-report-spans | test step failed 2-report-spans case.go:364: failed in step 2-report-spans case.go:366: --- Job:kuttl-test-cool-guinea/00-report-span +++ Job:kuttl-test-cool-guinea/00-report-span @@ -1,8 +1,162 @@ apiVersion: batch/v1 kind: Job metadata: + annotations: + kubectl.kubernetes.io/last-applied-configuration: | + {"apiVersion":"batch/v1","kind":"Job","metadata":{"annotations":{},"name":"00-report-span","namespace":"kuttl-test-cool-guinea"},"spec":{"backoffLimit":15,"template":{"spec":{"containers":[{"command":["./reporter","--days","5","--verbose"],"env":[{"name":"JAEGER_SERVICE_NAME","value":"smoke-test-service"},{"name":"OPERATION_NAME","value":"smoke-test-operation"},{"name":"JAEGER_ENDPOINT","value":"http://test-es-index-cleaner-with-prefix-collector-headless:14268/api/traces"},{"name":"JAEGER_QUERY","value":"https://test-es-index-cleaner-with-prefix-query/api/traces"},{"name":"SECRET_PATH","value":"/var/run/secrets/api-token/token"}],"image":"quay.io/rhn_support_ikanse/jaeger-asserts:latest","name":"asserts-container","volumeMounts":[{"mountPath":"/var/run/secrets/api-token","name":"token-api-volume"}]}],"restartPolicy":"OnFailure","volumes":[{"name":"token-api-volume","secret":{"secretName":"e2e-test"}}]}}}} + labels: + batch.kubernetes.io/controller-uid: 34f85c1b-e0d1-4a3c-b003-5666f25c506d + batch.kubernetes.io/job-name: 00-report-span + controller-uid: 34f85c1b-e0d1-4a3c-b003-5666f25c506d + job-name: 00-report-span + managedFields: + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:kubectl.kubernetes.io/last-applied-configuration: {} + f:spec: + f:backoffLimit: {} + f:completionMode: {} + f:completions: {} + f:manualSelector: {} + f:parallelism: {} + f:podReplacementPolicy: {} + f:suspend: {} + f:template: + f:spec: + f:containers: + k:{"name":"asserts-container"}: + .: {} + f:command: {} + f:env: + .: {} + k:{"name":"JAEGER_ENDPOINT"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"JAEGER_QUERY"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"JAEGER_SERVICE_NAME"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"OPERATION_NAME"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SECRET_PATH"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:name: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/var/run/secrets/api-token"}: + .: {} + f:mountPath: {} + f:name: {} + f:dnsPolicy: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"token-api-volume"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + manager: kubectl-client-side-apply + operation: Update + time: "2025-01-27T07:26:52Z" + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:status: + f:active: {} + f:ready: {} + f:startTime: {} + f:terminating: {} + f:uncountedTerminatedPods: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2025-01-27T07:26:55Z" name: 00-report-span namespace: kuttl-test-cool-guinea +spec: + backoffLimit: 15 + completionMode: NonIndexed + completions: 1 + manualSelector: false + parallelism: 1 + podReplacementPolicy: TerminatingOrFailed + selector: + matchLabels: + batch.kubernetes.io/controller-uid: 34f85c1b-e0d1-4a3c-b003-5666f25c506d + suspend: false + template: + metadata: + creationTimestamp: null + labels: + batch.kubernetes.io/controller-uid: 34f85c1b-e0d1-4a3c-b003-5666f25c506d + batch.kubernetes.io/job-name: 00-report-span + controller-uid: 34f85c1b-e0d1-4a3c-b003-5666f25c506d + job-name: 00-report-span + spec: + containers: + - command: + - ./reporter + - --days + - "5" + - --verbose + env: + - name: JAEGER_SERVICE_NAME + value: smoke-test-service + - name: OPERATION_NAME + value: smoke-test-operation + - name: JAEGER_ENDPOINT + value: http://test-es-index-cleaner-with-prefix-collector-headless:14268/api/traces + - name: JAEGER_QUERY + value: https://test-es-index-cleaner-with-prefix-query/api/traces + - name: SECRET_PATH + value: /var/run/secrets/api-token/token + image: quay.io/rhn_support_ikanse/jaeger-asserts:latest + imagePullPolicy: Always + name: asserts-container + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /var/run/secrets/api-token + name: token-api-volume + dnsPolicy: ClusterFirst + restartPolicy: OnFailure + schedulerName: default-scheduler + securityContext: {} + terminationGracePeriodSeconds: 30 + volumes: + - name: token-api-volume + secret: + defaultMode: 420 + secretName: e2e-test status: - succeeded: 1 + active: 1 + ready: 1 + startTime: "2025-01-27T07:26:52Z" + terminating: 0 + uncountedTerminatedPods: {} case.go:366: resource Job:kuttl-test-cool-guinea/00-report-span: .status.succeeded: key is missing from map logger.go:42: 07:36:52 | es-index-cleaner-autoprov | es-index-cleaner-autoprov events from ns kuttl-test-cool-guinea: logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:12 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestcoolguineatestesindexcleane-1-9c884dbc SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestcoolguineatestesindexcleane-1-9jmhpw replicaset-controller logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:12 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcoolguineatestesindexcleane-1-9jmhpw Binding Scheduled Successfully assigned kuttl-test-cool-guinea/elasticsearch-cdm-kuttltestcoolguineatestesindexcleane-1-9jmhpw to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:12 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcoolguineatestesindexcleane-1-9jmhpw FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:12 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestcoolguineatestesindexcleane-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestcoolguineatestesindexcleane-1-9c884dbc to 1 deployment-controller logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:13 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcoolguineatestesindexcleane-1-9jmhpw AddedInterface Add eth0 [10.129.2.22/23] from ovn-kubernetes multus logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:13 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcoolguineatestesindexcleane-1-9jmhpw.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:13 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcoolguineatestesindexcleane-1-9jmhpw.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:13 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcoolguineatestesindexcleane-1-9jmhpw.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:13 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcoolguineatestesindexcleane-1-9jmhpw.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:13 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcoolguineatestesindexcleane-1-9jmhpw.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:13 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcoolguineatestesindexcleane-1-9jmhpw.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:23 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcoolguineatestesindexcleane-1-9jmhpw.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:28 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcoolguineatestesindexcleane-1-9jmhpw.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:39 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-76c484975f-wmmwk Binding Scheduled Successfully assigned kuttl-test-cool-guinea/test-es-index-cleaner-with-prefix-collector-76c484975f-wmmwk to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:39 +0000 UTC Warning Pod test-es-index-cleaner-with-prefix-collector-76c484975f-wmmwk FailedMount MountVolume.SetUp failed for volume "test-es-index-cleaner-with-prefix-collector-tls-config-volume" : secret "test-es-index-cleaner-with-prefix-collector-headless-tls" not found kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:39 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-collector-76c484975f SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-collector-76c484975f-wmmwk replicaset-controller logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:39 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-collector ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-collector-76c484975f to 1 deployment-controller logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:39 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b9c9c75d9-dvlbv Binding Scheduled Successfully assigned kuttl-test-cool-guinea/test-es-index-cleaner-with-prefix-query-5b9c9c75d9-dvlbv to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:39 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-5b9c9c75d9 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-5b9c9c75d9-dvlbv replicaset-controller logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:39 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-5b9c9c75d9 to 1 deployment-controller logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:40 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-76c484975f-wmmwk AddedInterface Add eth0 [10.131.0.41/23] from ovn-kubernetes multus logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:40 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-76c484975f-wmmwk.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:40 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-76c484975f-wmmwk.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:40 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-76c484975f-wmmwk.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:40 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b9c9c75d9-dvlbv AddedInterface Add eth0 [10.128.2.42/23] from ovn-kubernetes multus logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:40 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b9c9c75d9-dvlbv.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:40 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b9c9c75d9-dvlbv.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:40 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b9c9c75d9-dvlbv.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:40 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b9c9c75d9-dvlbv.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:40 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b9c9c75d9-dvlbv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:40 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b9c9c75d9-dvlbv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:40 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b9c9c75d9-dvlbv.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:40 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b9c9c75d9-dvlbv.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:40 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b9c9c75d9-dvlbv.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:52 +0000 UTC Normal Pod 00-report-span-h6mt4 Binding Scheduled Successfully assigned kuttl-test-cool-guinea/00-report-span-h6mt4 to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:52 +0000 UTC Normal Pod 00-report-span-h6mt4 AddedInterface Add eth0 [10.131.0.42/23] from ovn-kubernetes multus logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:52 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-h6mt4 job-controller logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:53 +0000 UTC Normal Pod 00-report-span-h6mt4.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:53 +0000 UTC Normal Pod 00-report-span-h6mt4.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 950ms (950ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:54 +0000 UTC Normal Pod 00-report-span-h6mt4.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:54 +0000 UTC Normal Pod 00-report-span-h6mt4.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:54 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:54 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:26:54 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:27:39 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-76c484975f-wmmwk horizontal-pod-autoscaler logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:27:39 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-76c484975f-wmmwk horizontal-pod-autoscaler logger.go:42: 07:36:52 | es-index-cleaner-autoprov | 2025-01-27 07:27:39 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-76c484975f-wmmwk horizontal-pod-autoscaler logger.go:42: 07:36:52 | es-index-cleaner-autoprov | Deleting namespace: kuttl-test-cool-guinea === CONT kuttl/harness/es-from-aio-to-production logger.go:42: 07:37:04 | es-from-aio-to-production | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:37:04 | es-from-aio-to-production | Creating namespace: kuttl-test-chief-bear logger.go:42: 07:37:04 | es-from-aio-to-production/0-install | starting test step 0-install logger.go:42: 07:37:04 | es-from-aio-to-production/0-install | Jaeger:kuttl-test-chief-bear/my-jaeger created logger.go:42: 07:37:14 | es-from-aio-to-production/0-install | test step completed 0-install logger.go:42: 07:37:14 | es-from-aio-to-production/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:37:14 | es-from-aio-to-production/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:37:15 | es-from-aio-to-production/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:37:21 | es-from-aio-to-production/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:37:22 | es-from-aio-to-production/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:37:22 | es-from-aio-to-production/1-smoke-test | job.batch/report-span created logger.go:42: 07:37:22 | es-from-aio-to-production/1-smoke-test | job.batch/check-span created logger.go:42: 07:37:35 | es-from-aio-to-production/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:37:35 | es-from-aio-to-production/3-install | starting test step 3-install logger.go:42: 07:37:35 | es-from-aio-to-production/3-install | Jaeger:kuttl-test-chief-bear/my-jaeger updated logger.go:42: 07:38:11 | es-from-aio-to-production/3-install | test step completed 3-install logger.go:42: 07:38:11 | es-from-aio-to-production/4-smoke-test | starting test step 4-smoke-test logger.go:42: 07:38:11 | es-from-aio-to-production/4-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:38:19 | es-from-aio-to-production/4-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:38:19 | es-from-aio-to-production/4-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:38:19 | es-from-aio-to-production/4-smoke-test | job.batch/report-span unchanged logger.go:42: 07:38:19 | es-from-aio-to-production/4-smoke-test | job.batch/check-span unchanged logger.go:42: 07:38:19 | es-from-aio-to-production/4-smoke-test | test step completed 4-smoke-test logger.go:42: 07:38:19 | es-from-aio-to-production | es-from-aio-to-production events from ns kuttl-test-chief-bear: logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:08 +0000 UTC Normal Pod my-jaeger-85bccb7b4-hvsg2 Binding Scheduled Successfully assigned kuttl-test-chief-bear/my-jaeger-85bccb7b4-hvsg2 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:08 +0000 UTC Normal ReplicaSet.apps my-jaeger-85bccb7b4 SuccessfulCreate Created pod: my-jaeger-85bccb7b4-hvsg2 replicaset-controller logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:08 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-85bccb7b4 to 1 deployment-controller logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:09 +0000 UTC Normal Pod my-jaeger-85bccb7b4-hvsg2 AddedInterface Add eth0 [10.128.2.43/23] from ovn-kubernetes multus logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:09 +0000 UTC Normal Pod my-jaeger-85bccb7b4-hvsg2.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:12 +0000 UTC Normal Pod my-jaeger-85bccb7b4-hvsg2.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" in 3.533s (3.533s including waiting). Image size: 147411949 bytes. kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:12 +0000 UTC Normal Pod my-jaeger-85bccb7b4-hvsg2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:12 +0000 UTC Normal Pod my-jaeger-85bccb7b4-hvsg2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:12 +0000 UTC Normal Pod my-jaeger-85bccb7b4-hvsg2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:12 +0000 UTC Normal Pod my-jaeger-85bccb7b4-hvsg2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:12 +0000 UTC Normal Pod my-jaeger-85bccb7b4-hvsg2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:18 +0000 UTC Normal Pod my-jaeger-55ffb444f5-pq6qd Binding Scheduled Successfully assigned kuttl-test-chief-bear/my-jaeger-55ffb444f5-pq6qd to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:18 +0000 UTC Normal ReplicaSet.apps my-jaeger-55ffb444f5 SuccessfulCreate Created pod: my-jaeger-55ffb444f5-pq6qd replicaset-controller logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:18 +0000 UTC Normal Pod my-jaeger-85bccb7b4-hvsg2.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:18 +0000 UTC Normal Pod my-jaeger-85bccb7b4-hvsg2.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:18 +0000 UTC Normal ReplicaSet.apps my-jaeger-85bccb7b4 SuccessfulDelete Deleted pod: my-jaeger-85bccb7b4-hvsg2 replicaset-controller logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:18 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-85bccb7b4 to 0 from 1 deployment-controller logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:18 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-55ffb444f5 to 1 deployment-controller logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:19 +0000 UTC Normal Pod my-jaeger-55ffb444f5-pq6qd AddedInterface Add eth0 [10.128.2.44/23] from ovn-kubernetes multus logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:19 +0000 UTC Normal Pod my-jaeger-55ffb444f5-pq6qd.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:19 +0000 UTC Normal Pod my-jaeger-55ffb444f5-pq6qd.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:19 +0000 UTC Normal Pod my-jaeger-55ffb444f5-pq6qd.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:19 +0000 UTC Normal Pod my-jaeger-55ffb444f5-pq6qd.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:19 +0000 UTC Normal Pod my-jaeger-55ffb444f5-pq6qd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:19 +0000 UTC Normal Pod my-jaeger-55ffb444f5-pq6qd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:22 +0000 UTC Normal Pod check-span-hm64v Binding Scheduled Successfully assigned kuttl-test-chief-bear/check-span-hm64v to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:22 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-hm64v job-controller logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:22 +0000 UTC Normal Pod report-span-w485x Binding Scheduled Successfully assigned kuttl-test-chief-bear/report-span-w485x to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:22 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-w485x job-controller logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:23 +0000 UTC Normal Pod check-span-hm64v AddedInterface Add eth0 [10.131.0.44/23] from ovn-kubernetes multus logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:23 +0000 UTC Normal Pod check-span-hm64v.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:23 +0000 UTC Normal Pod report-span-w485x AddedInterface Add eth0 [10.129.2.23/23] from ovn-kubernetes multus logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:23 +0000 UTC Normal Pod report-span-w485x.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:24 +0000 UTC Normal Pod check-span-hm64v.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 696ms (696ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:24 +0000 UTC Normal Pod check-span-hm64v.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:24 +0000 UTC Normal Pod check-span-hm64v.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:25 +0000 UTC Normal Pod report-span-w485x.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2s (2s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:25 +0000 UTC Normal Pod report-span-w485x.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:25 +0000 UTC Normal Pod report-span-w485x.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:34 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchiefbearmyjaeger-1-77f96f8c97-fq8j2 Binding Scheduled Successfully assigned kuttl-test-chief-bear/elasticsearch-cdm-kuttltestchiefbearmyjaeger-1-77f96f8c97-fq8j2 to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchiefbearmyjaeger-1-77f96f8c97-fq8j2 AddedInterface Add eth0 [10.131.0.45/23] from ovn-kubernetes multus logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchiefbearmyjaeger-1-77f96f8c97-fq8j2.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchiefbearmyjaeger-1-77f96f8c97-fq8j2.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchiefbearmyjaeger-1-77f96f8c97-fq8j2.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchiefbearmyjaeger-1-77f96f8c97-fq8j2.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchiefbearmyjaeger-1-77f96f8c97-fq8j2.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchiefbearmyjaeger-1-77f96f8c97-fq8j2.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:39 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestchiefbearmyjaeger-1-77f96f8c97 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestchiefbearmyjaeger-1-77f96f8c97-fq8j2 replicaset-controller logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:39 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestchiefbearmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestchiefbearmyjaeger-1-77f96f8c97 to 1 deployment-controller logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:49 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestchiefbearmyjaeger-1-77f96f8c97-fq8j2.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:54 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestchiefbearmyjaeger-1-77f96f8c97-fq8j2.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:37:57 +0000 UTC Normal Job.batch report-span Completed Job completed job-controller logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:06 +0000 UTC Normal Pod my-jaeger-collector-6bd97fd4c4-pk492 Binding Scheduled Successfully assigned kuttl-test-chief-bear/my-jaeger-collector-6bd97fd4c4-pk492 to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:06 +0000 UTC Normal Pod my-jaeger-collector-6bd97fd4c4-pk492 AddedInterface Add eth0 [10.129.2.24/23] from ovn-kubernetes multus logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:06 +0000 UTC Normal Pod my-jaeger-collector-6bd97fd4c4-pk492.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:06 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-6bd97fd4c4 SuccessfulCreate Created pod: my-jaeger-collector-6bd97fd4c4-pk492 replicaset-controller logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:06 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-6bd97fd4c4 to 1 deployment-controller logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:06 +0000 UTC Normal Pod my-jaeger-query-748ccf8bc7-zwh2s Binding Scheduled Successfully assigned kuttl-test-chief-bear/my-jaeger-query-748ccf8bc7-zwh2s to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:06 +0000 UTC Normal Pod my-jaeger-query-748ccf8bc7-zwh2s AddedInterface Add eth0 [10.128.2.45/23] from ovn-kubernetes multus logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:06 +0000 UTC Normal Pod my-jaeger-query-748ccf8bc7-zwh2s.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:06 +0000 UTC Normal Pod my-jaeger-query-748ccf8bc7-zwh2s.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:06 +0000 UTC Normal Pod my-jaeger-query-748ccf8bc7-zwh2s.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:06 +0000 UTC Normal Pod my-jaeger-query-748ccf8bc7-zwh2s.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:06 +0000 UTC Normal Pod my-jaeger-query-748ccf8bc7-zwh2s.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:06 +0000 UTC Normal Pod my-jaeger-query-748ccf8bc7-zwh2s.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:06 +0000 UTC Normal Pod my-jaeger-query-748ccf8bc7-zwh2s.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:06 +0000 UTC Normal Pod my-jaeger-query-748ccf8bc7-zwh2s.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:06 +0000 UTC Normal Pod my-jaeger-query-748ccf8bc7-zwh2s.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:06 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-748ccf8bc7 SuccessfulCreate Created pod: my-jaeger-query-748ccf8bc7-zwh2s replicaset-controller logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:06 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-748ccf8bc7 to 1 deployment-controller logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:08 +0000 UTC Normal Pod my-jaeger-55ffb444f5-pq6qd.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:08 +0000 UTC Normal Pod my-jaeger-55ffb444f5-pq6qd.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:10 +0000 UTC Normal Pod my-jaeger-collector-6bd97fd4c4-pk492.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" in 3.697s (3.697s including waiting). Image size: 142489237 bytes. kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:10 +0000 UTC Normal Pod my-jaeger-collector-6bd97fd4c4-pk492.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | 2025-01-27 07:38:10 +0000 UTC Normal Pod my-jaeger-collector-6bd97fd4c4-pk492.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:38:19 | es-from-aio-to-production | Deleting namespace: kuttl-test-chief-bear === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (1204.35s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.96s) --- PASS: kuttl/harness/es-multiinstance (119.77s) --- PASS: kuttl/harness/es-simple-prod (6.00s) --- PASS: kuttl/harness/es-rollover-autoprov (227.43s) --- PASS: kuttl/harness/es-increasing-replicas (105.53s) --- FAIL: kuttl/harness/es-index-cleaner-autoprov (657.97s) --- PASS: kuttl/harness/es-from-aio-to-production (81.65s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name elasticsearch --report --output /logs/artifacts/elasticsearch.xml ./artifacts/kuttl-report.xml time="2025-01-27T07:38:27Z" level=debug msg="Setting a new name for the test suites" time="2025-01-27T07:38:27Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-01-27T07:38:27Z" level=debug msg="normalizing test case names" time="2025-01-27T07:38:27Z" level=debug msg="elasticsearch/artifacts -> elasticsearch_artifacts" time="2025-01-27T07:38:27Z" level=debug msg="elasticsearch/es-multiinstance -> elasticsearch_es_multiinstance" time="2025-01-27T07:38:27Z" level=debug msg="elasticsearch/es-simple-prod -> elasticsearch_es_simple_prod" time="2025-01-27T07:38:27Z" level=debug msg="elasticsearch/es-rollover-autoprov -> elasticsearch_es_rollover_autoprov" time="2025-01-27T07:38:27Z" level=debug msg="elasticsearch/es-increasing-replicas -> elasticsearch_es_increasing_replicas" time="2025-01-27T07:38:27Z" level=debug msg="elasticsearch/es-index-cleaner-autoprov -> elasticsearch_es_index_cleaner_autoprov" time="2025-01-27T07:38:27Z" level=debug msg="elasticsearch/es-from-aio-to-production -> elasticsearch_es_from_aio_to_production" +-----------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------+--------+ | elasticsearch_artifacts | passed | | elasticsearch_es_multiinstance | passed | | elasticsearch_es_simple_prod | passed | | elasticsearch_es_rollover_autoprov | passed | | elasticsearch_es_increasing_replicas | passed | | elasticsearch_es_index_cleaner_autoprov | failed | | elasticsearch_es_from_aio_to_production | passed | +-----------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh examples false true + '[' 3 -ne 3 ']' + test_suite_name=examples + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/examples.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-examples make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ VERTX_IMG=jaegertracing/vertx-create-span:operator-e2e-tests \ ./tests/e2e/examples/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-25-163410 True False 27m Cluster version is 4.18.0-0.nightly-2025-01-25-163410' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-25-163410 True False 27m Cluster version is 4.18.0-0.nightly-2025-01-25-163410' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/examples/render.sh ++ export SUITE_DIR=./tests/e2e/examples ++ SUITE_DIR=./tests/e2e/examples ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/examples ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test examples-agent-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-agent-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-with-priority-class\e[0m' Rendering files for test examples-agent-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + mkdir -p examples-agent-with-priority-class + cd examples-agent-with-priority-class + example_name=agent-with-priority-class + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + '[' true '!=' true ']' + render_install_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + install_file=./02-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/agent-with-priority-class.yaml -o ./02-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./02-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./02-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./02-install.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + JAEGER_NAME=agent-as-daemonset + local jaeger_strategy ++ get_jaeger_strategy ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./02-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./02-install.yaml ++ strategy=DaemonSet ++ '[' DaemonSet = null ']' ++ echo DaemonSet ++ return 0 + jaeger_strategy=DaemonSet + '[' DaemonSet = DaemonSet ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./02-assert.yaml + render_smoke_test_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + jaeger_name=agent-as-daemonset + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test agent-as-daemonset true 02 + '[' 3 -ne 3 ']' + jaeger=agent-as-daemonset + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-all-in-one-with-options + '[' 1 -ne 1 ']' + test_name=examples-all-in-one-with-options + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-all-in-one-with-options' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-all-in-one-with-options\e[0m' Rendering files for test examples-all-in-one-with-options + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-with-priority-class + '[' examples-agent-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-all-in-one-with-options + cd examples-all-in-one-with-options + example_name=all-in-one-with-options + render_install_example all-in-one-with-options 00 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/all-in-one-with-options.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + JAEGER_NAME=my-jaeger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.metadata.name="my-jaeger"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i 'del(.spec.allInOne.image)' ./00-install.yaml + render_smoke_test_example all-in-one-with-options 01 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + jaeger_name=my-jaeger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + sed -i s~my-jaeger-query:443~my-jaeger-query:443/jaeger~gi ./01-smoke-test.yaml + start_test examples-business-application-injected-sidecar + '[' 1 -ne 1 ']' + test_name=examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-business-application-injected-sidecar' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-business-application-injected-sidecar\e[0m' Rendering files for test examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-all-in-one-with-options + '[' examples-all-in-one-with-options '!=' _build ']' + cd .. + mkdir -p examples-business-application-injected-sidecar + cd examples-business-application-injected-sidecar + example_name=simplest + cp /tmp/jaeger-tests/examples/business-application-injected-sidecar.yaml ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].image=strenv(VERTX_IMG)' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.path="/"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.port=8080' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.initialDelaySeconds=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.failureThreshold=3' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.periodSeconds=10' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.successThreshold=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.timeoutSeconds=1' ./00-install.yaml + render_install_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example simplest 02 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 02 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-collector-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-collector-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-collector-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-collector-with-priority-class\e[0m' Rendering files for test examples-collector-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-business-application-injected-sidecar + '[' examples-business-application-injected-sidecar '!=' _build ']' + cd .. + mkdir -p examples-collector-with-priority-class + cd examples-collector-with-priority-class + example_name=collector-with-priority-class + render_install_example collector-with-priority-class 00 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/collector-with-priority-class.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + JAEGER_NAME=collector-with-high-priority + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example collector-with-priority-class 01 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + jaeger_name=collector-with-high-priority + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test collector-with-high-priority true 01 + '[' 3 -ne 3 ']' + jaeger=collector-with-high-priority + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + export JAEGER_NAME=collector-with-high-priority + JAEGER_NAME=collector-with-high-priority + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-service-types + '[' 1 -ne 1 ']' + test_name=examples-service-types + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-service-types' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-service-types\e[0m' Rendering files for test examples-service-types + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-collector-with-priority-class + '[' examples-collector-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-service-types + cd examples-service-types + example_name=service-types + render_install_example service-types 00 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/service-types.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + JAEGER_NAME=service-types + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example service-types 01 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/service-types.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/service-types.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/service-types.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/service-types.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + jaeger_name=service-types + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test service-types true 01 + '[' 3 -ne 3 ']' + jaeger=service-types + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + export JAEGER_NAME=service-types + JAEGER_NAME=service-types + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod + '[' 1 -ne 1 ']' + test_name=examples-simple-prod + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod\e[0m' Rendering files for test examples-simple-prod + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-service-types + '[' examples-service-types '!=' _build ']' + cd .. + mkdir -p examples-simple-prod + cd examples-simple-prod + example_name=simple-prod + render_install_example simple-prod 01 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod 02 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod-with-volumes + '[' 1 -ne 1 ']' + test_name=examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod-with-volumes' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod-with-volumes\e[0m' Rendering files for test examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod + '[' examples-simple-prod '!=' _build ']' + cd .. + mkdir -p examples-simple-prod-with-volumes + cd examples-simple-prod-with-volumes + example_name=simple-prod-with-volumes + render_install_example simple-prod-with-volumes 01 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod-with-volumes 02 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + /tmp/jaeger-tests/bin/gomplate -f ./03-check-volume.yaml.template -o 03-check-volume.yaml + start_test examples-simplest + '[' 1 -ne 1 ']' + test_name=examples-simplest + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simplest' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simplest\e[0m' Rendering files for test examples-simplest + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod-with-volumes + '[' examples-simple-prod-with-volumes '!=' _build ']' + cd .. + mkdir -p examples-simplest + cd examples-simplest + example_name=simplest + render_install_example simplest 00 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 01 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger + '[' 1 -ne 1 ']' + test_name=examples-with-badger + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger\e[0m' Rendering files for test examples-with-badger + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simplest + '[' examples-simplest '!=' _build ']' + cd .. + mkdir -p examples-with-badger + cd examples-with-badger + example_name=with-badger + render_install_example with-badger 00 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + JAEGER_NAME=with-badger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger 01 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + jaeger_name=with-badger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + export JAEGER_NAME=with-badger + JAEGER_NAME=with-badger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger-and-volume + '[' 1 -ne 1 ']' + test_name=examples-with-badger-and-volume + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger-and-volume' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger-and-volume\e[0m' Rendering files for test examples-with-badger-and-volume + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger + '[' examples-with-badger '!=' _build ']' + cd .. + mkdir -p examples-with-badger-and-volume + cd examples-with-badger-and-volume + example_name=with-badger-and-volume + render_install_example with-badger-and-volume 00 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger-and-volume.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + JAEGER_NAME=with-badger-and-volume + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger-and-volume 01 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + jaeger_name=with-badger-and-volume + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger-and-volume true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger-and-volume + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + export JAEGER_NAME=with-badger-and-volume + JAEGER_NAME=with-badger-and-volume + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-cassandra + '[' 1 -ne 1 ']' + test_name=examples-with-cassandra + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-cassandra' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-cassandra\e[0m' Rendering files for test examples-with-cassandra + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger-and-volume + '[' examples-with-badger-and-volume '!=' _build ']' + cd .. + mkdir -p examples-with-cassandra + cd examples-with-cassandra + example_name=with-cassandra + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-cassandra 01 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-cassandra.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + JAEGER_NAME=with-cassandra + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-cassandra 02 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-cassandra.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-cassandra.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + jaeger_name=with-cassandra + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-cassandra true 02 + '[' 3 -ne 3 ']' + jaeger=with-cassandra + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + export JAEGER_NAME=with-cassandra + JAEGER_NAME=with-cassandra + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-sampling + '[' 1 -ne 1 ']' + test_name=examples-with-sampling + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-sampling' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-sampling\e[0m' Rendering files for test examples-with-sampling + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-cassandra + '[' examples-with-cassandra '!=' _build ']' + cd .. + mkdir -p examples-with-sampling + cd examples-with-sampling + export example_name=with-sampling + example_name=with-sampling + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-sampling 01 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-sampling.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + JAEGER_NAME=with-sampling + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-sampling 02 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-sampling.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-sampling.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + jaeger_name=with-sampling + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-sampling true 02 + '[' 3 -ne 3 ']' + jaeger=with-sampling + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + export JAEGER_NAME=with-sampling + JAEGER_NAME=with-sampling + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-agent-as-daemonset + '[' 1 -ne 1 ']' + test_name=examples-agent-as-daemonset + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-as-daemonset' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-as-daemonset\e[0m' Rendering files for test examples-agent-as-daemonset + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-sampling + '[' examples-with-sampling '!=' _build ']' + cd .. + mkdir -p examples-agent-as-daemonset + cd examples-agent-as-daemonset + '[' true = true ']' + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/agent-as-daemonset.yaml -o 02-install.yaml + '[' true = true ']' + start_test examples-openshift-with-htpasswd + '[' 1 -ne 1 ']' + test_name=examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-openshift-with-htpasswd' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-openshift-with-htpasswd\e[0m' Rendering files for test examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-as-daemonset + '[' examples-agent-as-daemonset '!=' _build ']' + cd .. + mkdir -p examples-openshift-with-htpasswd + cd examples-openshift-with-htpasswd + export JAEGER_NAME=with-htpasswd + JAEGER_NAME=with-htpasswd + export JAEGER_USERNAME=awesomeuser + JAEGER_USERNAME=awesomeuser + export JAEGER_PASSWORD=awesomepassword + JAEGER_PASSWORD=awesomepassword + export 'JAEGER_USER_PASSWORD_HASH=awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' + JAEGER_USER_PASSWORD_HASH='awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ echo 'awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ base64 + SECRET=YXdlc29tZXVzZXI6e1NIQX11VWRxUFZVeXFOQm1FUlUwUXhqM0tGYVpuanc9Cg== + /tmp/jaeger-tests/bin/gomplate -f ./00-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/with-htpasswd.yaml -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + export 'GET_URL_COMMAND=kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + export 'URL=https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + INSECURE=true + JAEGER_USERNAME= + JAEGER_PASSWORD= + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-unsecured.yaml + JAEGER_USERNAME=wronguser + JAEGER_PASSWORD=wrongpassword + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-check-unauthorized.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./04-check-authorized.yaml + skip_test examples-agent-as-daemonset 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-agent-as-daemonset + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-openshift-with-htpasswd + '[' examples-openshift-with-htpasswd '!=' _build ']' + cd .. + rm -rf examples-agent-as-daemonset + warning 'examples-agent-as-daemonset: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-agent-as-daemonset: This test is flaky in Prow CI\e[0m' WAR: examples-agent-as-daemonset: This test is flaky in Prow CI + skip_test examples-with-badger-and-volume 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-with-badger-and-volume + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-with-badger-and-volume + warning 'examples-with-badger-and-volume: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-with-badger-and-volume: This test is flaky in Prow CI\e[0m' WAR: examples-with-badger-and-volume: This test is flaky in Prow CI + skip_test examples-collector-with-priority-class 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-collector-with-priority-class + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-collector-with-priority-class + warning 'examples-collector-with-priority-class: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-collector-with-priority-class: This test is flaky in Prow CI\e[0m' WAR: examples-collector-with-priority-class: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running examples E2E tests' Running examples E2E tests + cd tests/e2e/examples/_build + set +e + KUBECONFIG=/tmp/kubeconfig-4053090600 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 12 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/examples-agent-with-priority-class === PAUSE kuttl/harness/examples-agent-with-priority-class === RUN kuttl/harness/examples-all-in-one-with-options === PAUSE kuttl/harness/examples-all-in-one-with-options === RUN kuttl/harness/examples-business-application-injected-sidecar === PAUSE kuttl/harness/examples-business-application-injected-sidecar === RUN kuttl/harness/examples-openshift-with-htpasswd === PAUSE kuttl/harness/examples-openshift-with-htpasswd === RUN kuttl/harness/examples-service-types === PAUSE kuttl/harness/examples-service-types === RUN kuttl/harness/examples-simple-prod === PAUSE kuttl/harness/examples-simple-prod === RUN kuttl/harness/examples-simple-prod-with-volumes === PAUSE kuttl/harness/examples-simple-prod-with-volumes === RUN kuttl/harness/examples-simplest === PAUSE kuttl/harness/examples-simplest === RUN kuttl/harness/examples-with-badger === PAUSE kuttl/harness/examples-with-badger === RUN kuttl/harness/examples-with-cassandra === PAUSE kuttl/harness/examples-with-cassandra === RUN kuttl/harness/examples-with-sampling === PAUSE kuttl/harness/examples-with-sampling === CONT kuttl/harness/artifacts logger.go:42: 07:38:57 | artifacts | Creating namespace: kuttl-test-on-terrapin logger.go:42: 07:38:57 | artifacts | artifacts events from ns kuttl-test-on-terrapin: logger.go:42: 07:38:57 | artifacts | Deleting namespace: kuttl-test-on-terrapin === CONT kuttl/harness/examples-simple-prod logger.go:42: 07:39:03 | examples-simple-prod | Creating namespace: kuttl-test-holy-imp logger.go:42: 07:39:03 | examples-simple-prod/1-install | starting test step 1-install logger.go:42: 07:39:03 | examples-simple-prod/1-install | Jaeger:kuttl-test-holy-imp/simple-prod created logger.go:42: 07:39:40 | examples-simple-prod/1-install | test step completed 1-install logger.go:42: 07:39:40 | examples-simple-prod/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:39:40 | examples-simple-prod/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:39:41 | examples-simple-prod/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:39:50 | examples-simple-prod/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:39:50 | examples-simple-prod/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:39:50 | examples-simple-prod/2-smoke-test | job.batch/report-span created logger.go:42: 07:39:50 | examples-simple-prod/2-smoke-test | job.batch/check-span created logger.go:42: 07:39:59 | examples-simple-prod/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:39:59 | examples-simple-prod | examples-simple-prod events from ns kuttl-test-holy-imp: logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestholyimpsimpleprod-1-77b4ff4849-hg4b7 Binding Scheduled Successfully assigned kuttl-test-holy-imp/elasticsearch-cdm-kuttltestholyimpsimpleprod-1-77b4ff4849-hg4b7 to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:10 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestholyimpsimpleprod-1-77b4ff4849 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestholyimpsimpleprod-1-77b4ff4849-hg4b7 replicaset-controller logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:10 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestholyimpsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestholyimpsimpleprod-1-77b4ff4849 to 1 deployment-controller logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestholyimpsimpleprod-1-77b4ff4849-hg4b7 AddedInterface Add eth0 [10.131.0.46/23] from ovn-kubernetes multus logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestholyimpsimpleprod-1-77b4ff4849-hg4b7.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestholyimpsimpleprod-1-77b4ff4849-hg4b7.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestholyimpsimpleprod-1-77b4ff4849-hg4b7.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestholyimpsimpleprod-1-77b4ff4849-hg4b7.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestholyimpsimpleprod-1-77b4ff4849-hg4b7.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestholyimpsimpleprod-1-77b4ff4849-hg4b7.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:25 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestholyimpsimpleprod-1-77b4ff4849-hg4b7.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:37 +0000 UTC Normal Pod simple-prod-collector-7648dbfff7-r6v7b Binding Scheduled Successfully assigned kuttl-test-holy-imp/simple-prod-collector-7648dbfff7-r6v7b to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:37 +0000 UTC Normal Pod simple-prod-collector-7648dbfff7-r6v7b AddedInterface Add eth0 [10.129.2.25/23] from ovn-kubernetes multus logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:37 +0000 UTC Normal Pod simple-prod-collector-7648dbfff7-r6v7b.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:37 +0000 UTC Normal Pod simple-prod-collector-7648dbfff7-r6v7b.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:37 +0000 UTC Normal Pod simple-prod-collector-7648dbfff7-r6v7b.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:37 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-7648dbfff7 SuccessfulCreate Created pod: simple-prod-collector-7648dbfff7-r6v7b replicaset-controller logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:37 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-7648dbfff7 to 1 deployment-controller logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:37 +0000 UTC Normal Pod simple-prod-query-78c8f57bcb-jgr9f Binding Scheduled Successfully assigned kuttl-test-holy-imp/simple-prod-query-78c8f57bcb-jgr9f to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:37 +0000 UTC Normal ReplicaSet.apps simple-prod-query-78c8f57bcb SuccessfulCreate Created pod: simple-prod-query-78c8f57bcb-jgr9f replicaset-controller logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:37 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-78c8f57bcb to 1 deployment-controller logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:38 +0000 UTC Warning Pod simple-prod-query-78c8f57bcb-jgr9f FailedMount MountVolume.SetUp failed for volume "simple-prod-ui-configuration-volume" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:38 +0000 UTC Warning Pod simple-prod-query-78c8f57bcb-jgr9f FailedMount MountVolume.SetUp failed for volume "simple-prod-service-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:38 +0000 UTC Warning Pod simple-prod-query-78c8f57bcb-jgr9f FailedMount MountVolume.SetUp failed for volume "certs" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:38 +0000 UTC Warning Pod simple-prod-query-78c8f57bcb-jgr9f FailedMount MountVolume.SetUp failed for volume "simple-prod-ui-oauth-proxy-tls" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:39 +0000 UTC Normal Pod simple-prod-query-78c8f57bcb-jgr9f AddedInterface Add eth0 [10.128.2.46/23] from ovn-kubernetes multus logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:39 +0000 UTC Normal Pod simple-prod-query-78c8f57bcb-jgr9f.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:39 +0000 UTC Normal Pod simple-prod-query-78c8f57bcb-jgr9f.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:39 +0000 UTC Normal Pod simple-prod-query-78c8f57bcb-jgr9f.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:39 +0000 UTC Normal Pod simple-prod-query-78c8f57bcb-jgr9f.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:39 +0000 UTC Normal Pod simple-prod-query-78c8f57bcb-jgr9f.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:39 +0000 UTC Normal Pod simple-prod-query-78c8f57bcb-jgr9f.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:39 +0000 UTC Normal Pod simple-prod-query-78c8f57bcb-jgr9f.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:39 +0000 UTC Normal Pod simple-prod-query-78c8f57bcb-jgr9f.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:39 +0000 UTC Normal Pod simple-prod-query-78c8f57bcb-jgr9f.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:47 +0000 UTC Normal Pod simple-prod-query-78c8f57bcb-jgr9f.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:47 +0000 UTC Normal Pod simple-prod-query-78c8f57bcb-jgr9f.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:47 +0000 UTC Normal Pod simple-prod-query-78c8f57bcb-jgr9f.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:47 +0000 UTC Normal ReplicaSet.apps simple-prod-query-78c8f57bcb SuccessfulDelete Deleted pod: simple-prod-query-78c8f57bcb-jgr9f replicaset-controller logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:47 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-78c8f57bcb to 0 from 1 deployment-controller logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:48 +0000 UTC Normal Pod simple-prod-query-868cd59559-ndm8k Binding Scheduled Successfully assigned kuttl-test-holy-imp/simple-prod-query-868cd59559-ndm8k to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:48 +0000 UTC Normal Pod simple-prod-query-868cd59559-ndm8k AddedInterface Add eth0 [10.128.2.47/23] from ovn-kubernetes multus logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:48 +0000 UTC Normal Pod simple-prod-query-868cd59559-ndm8k.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:48 +0000 UTC Normal Pod simple-prod-query-868cd59559-ndm8k.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:48 +0000 UTC Normal Pod simple-prod-query-868cd59559-ndm8k.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:48 +0000 UTC Normal Pod simple-prod-query-868cd59559-ndm8k.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:48 +0000 UTC Normal Pod simple-prod-query-868cd59559-ndm8k.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:48 +0000 UTC Normal Pod simple-prod-query-868cd59559-ndm8k.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:48 +0000 UTC Normal Pod simple-prod-query-868cd59559-ndm8k.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:48 +0000 UTC Normal Pod simple-prod-query-868cd59559-ndm8k.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:48 +0000 UTC Normal Pod simple-prod-query-868cd59559-ndm8k.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:48 +0000 UTC Normal ReplicaSet.apps simple-prod-query-868cd59559 SuccessfulCreate Created pod: simple-prod-query-868cd59559-ndm8k replicaset-controller logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:48 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-868cd59559 to 1 deployment-controller logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:50 +0000 UTC Normal Pod check-span-wrwkm Binding Scheduled Successfully assigned kuttl-test-holy-imp/check-span-wrwkm to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:50 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-wrwkm job-controller logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:50 +0000 UTC Normal Pod report-span-vcmxc Binding Scheduled Successfully assigned kuttl-test-holy-imp/report-span-vcmxc to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:50 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-vcmxc job-controller logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:51 +0000 UTC Normal Pod check-span-wrwkm AddedInterface Add eth0 [10.129.2.27/23] from ovn-kubernetes multus logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:51 +0000 UTC Normal Pod check-span-wrwkm.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:51 +0000 UTC Normal Pod report-span-vcmxc AddedInterface Add eth0 [10.129.2.26/23] from ovn-kubernetes multus logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:51 +0000 UTC Normal Pod report-span-vcmxc.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:52 +0000 UTC Normal Pod report-span-vcmxc.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.048s (1.049s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:52 +0000 UTC Normal Pod report-span-vcmxc.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:52 +0000 UTC Normal Pod report-span-vcmxc.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:52 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:52 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:52 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:55 +0000 UTC Normal Pod check-span-wrwkm.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 4.604s (4.604s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:56 +0000 UTC Normal Pod check-span-wrwkm.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:56 +0000 UTC Normal Pod check-span-wrwkm.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:39:59 | examples-simple-prod | 2025-01-27 07:39:58 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:39:59 | examples-simple-prod | Deleting namespace: kuttl-test-holy-imp === CONT kuttl/harness/examples-with-sampling logger.go:42: 07:40:12 | examples-with-sampling | Creating namespace: kuttl-test-social-spaniel logger.go:42: 07:40:12 | examples-with-sampling/0-install | starting test step 0-install logger.go:42: 07:40:12 | examples-with-sampling/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:40:12 | examples-with-sampling/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:40:12 | examples-with-sampling/0-install | >>>> Creating namespace kuttl-test-social-spaniel logger.go:42: 07:40:12 | examples-with-sampling/0-install | kubectl create namespace kuttl-test-social-spaniel 2>&1 | grep -v "already exists" || true logger.go:42: 07:40:12 | examples-with-sampling/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-social-spaniel 2>&1 | grep -v "already exists" || true logger.go:42: 07:40:12 | examples-with-sampling/0-install | service/cassandra created logger.go:42: 07:40:12 | examples-with-sampling/0-install | statefulset.apps/cassandra created logger.go:42: 07:40:12 | examples-with-sampling/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:40:25 | examples-with-sampling/0-install | test step completed 0-install logger.go:42: 07:40:25 | examples-with-sampling/1-install | starting test step 1-install logger.go:42: 07:40:25 | examples-with-sampling/1-install | Jaeger:kuttl-test-social-spaniel/with-sampling created logger.go:42: 07:40:31 | examples-with-sampling/1-install | test step completed 1-install logger.go:42: 07:40:31 | examples-with-sampling/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:40:31 | examples-with-sampling/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-sampling /dev/null] logger.go:42: 07:40:32 | examples-with-sampling/2-smoke-test | Warning: resource jaegers/with-sampling is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:40:39 | examples-with-sampling/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:40:39 | examples-with-sampling/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:40:39 | examples-with-sampling/2-smoke-test | job.batch/report-span created logger.go:42: 07:40:39 | examples-with-sampling/2-smoke-test | job.batch/check-span created logger.go:42: 07:40:51 | examples-with-sampling/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:40:51 | examples-with-sampling/3- | starting test step 3- logger.go:42: 07:40:51 | examples-with-sampling/3- | test step completed 3- logger.go:42: 07:40:51 | examples-with-sampling | examples-with-sampling events from ns kuttl-test-social-spaniel: logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:12 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-social-spaniel/cassandra-0 to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:12 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:13 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.129.2.28/23] from ovn-kubernetes multus logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:13 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:20 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 7.348s (7.348s including waiting). Image size: 309824447 bytes. kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:20 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:20 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:20 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-social-spaniel/cassandra-1 to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:20 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:21 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.131.0.47/23] from ovn-kubernetes multus logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:21 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:25 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 3.722s (3.722s including waiting). Image size: 309824447 bytes. kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:25 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:25 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:29 +0000 UTC Normal Pod with-sampling-bd9765b69-kklwd Binding Scheduled Successfully assigned kuttl-test-social-spaniel/with-sampling-bd9765b69-kklwd to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:29 +0000 UTC Normal ReplicaSet.apps with-sampling-bd9765b69 SuccessfulCreate Created pod: with-sampling-bd9765b69-kklwd replicaset-controller logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:29 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-bd9765b69 to 1 deployment-controller logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:30 +0000 UTC Normal Pod with-sampling-bd9765b69-kklwd AddedInterface Add eth0 [10.128.2.48/23] from ovn-kubernetes multus logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:30 +0000 UTC Normal Pod with-sampling-bd9765b69-kklwd.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:30 +0000 UTC Normal Pod with-sampling-bd9765b69-kklwd.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:30 +0000 UTC Normal Pod with-sampling-bd9765b69-kklwd.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:30 +0000 UTC Normal Pod with-sampling-bd9765b69-kklwd.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:30 +0000 UTC Normal Pod with-sampling-bd9765b69-kklwd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:30 +0000 UTC Normal Pod with-sampling-bd9765b69-kklwd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:37 +0000 UTC Normal Pod with-sampling-bd9765b69-kklwd.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:37 +0000 UTC Normal Pod with-sampling-bd9765b69-kklwd.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:37 +0000 UTC Normal ReplicaSet.apps with-sampling-bd9765b69 SuccessfulDelete Deleted pod: with-sampling-bd9765b69-kklwd replicaset-controller logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:37 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled down replica set with-sampling-bd9765b69 to 0 from 1 deployment-controller logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:38 +0000 UTC Normal Pod with-sampling-b54f64d78-fdsjf Binding Scheduled Successfully assigned kuttl-test-social-spaniel/with-sampling-b54f64d78-fdsjf to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:38 +0000 UTC Normal Pod with-sampling-b54f64d78-fdsjf AddedInterface Add eth0 [10.128.2.49/23] from ovn-kubernetes multus logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:38 +0000 UTC Normal Pod with-sampling-b54f64d78-fdsjf.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:38 +0000 UTC Normal Pod with-sampling-b54f64d78-fdsjf.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:38 +0000 UTC Normal Pod with-sampling-b54f64d78-fdsjf.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:38 +0000 UTC Normal Pod with-sampling-b54f64d78-fdsjf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:38 +0000 UTC Normal Pod with-sampling-b54f64d78-fdsjf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:38 +0000 UTC Normal Pod with-sampling-b54f64d78-fdsjf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:38 +0000 UTC Normal ReplicaSet.apps with-sampling-b54f64d78 SuccessfulCreate Created pod: with-sampling-b54f64d78-fdsjf replicaset-controller logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:38 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-b54f64d78 to 1 deployment-controller logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:39 +0000 UTC Normal Pod check-span-jwrpd Binding Scheduled Successfully assigned kuttl-test-social-spaniel/check-span-jwrpd to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:39 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-jwrpd job-controller logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:39 +0000 UTC Normal Pod report-span-j5jn8 Binding Scheduled Successfully assigned kuttl-test-social-spaniel/report-span-j5jn8 to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:39 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-j5jn8 job-controller logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:40 +0000 UTC Normal Pod check-span-jwrpd AddedInterface Add eth0 [10.131.0.48/23] from ovn-kubernetes multus logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:40 +0000 UTC Normal Pod check-span-jwrpd.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:40 +0000 UTC Normal Pod check-span-jwrpd.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 521ms (521ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:40 +0000 UTC Normal Pod report-span-j5jn8 AddedInterface Add eth0 [10.129.2.29/23] from ovn-kubernetes multus logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:40 +0000 UTC Normal Pod report-span-j5jn8.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:41 +0000 UTC Normal Pod check-span-jwrpd.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:41 +0000 UTC Normal Pod check-span-jwrpd.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:41 +0000 UTC Normal Pod report-span-j5jn8.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 636ms (636ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:41 +0000 UTC Normal Pod report-span-j5jn8.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:41 +0000 UTC Normal Pod report-span-j5jn8.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:40:51 | examples-with-sampling | 2025-01-27 07:40:51 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:40:51 | examples-with-sampling | Deleting namespace: kuttl-test-social-spaniel === CONT kuttl/harness/examples-with-cassandra logger.go:42: 07:42:01 | examples-with-cassandra | Creating namespace: kuttl-test-harmless-tiger logger.go:42: 07:42:01 | examples-with-cassandra/0-install | starting test step 0-install logger.go:42: 07:42:01 | examples-with-cassandra/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:42:01 | examples-with-cassandra/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:42:01 | examples-with-cassandra/0-install | >>>> Creating namespace kuttl-test-harmless-tiger logger.go:42: 07:42:01 | examples-with-cassandra/0-install | kubectl create namespace kuttl-test-harmless-tiger 2>&1 | grep -v "already exists" || true logger.go:42: 07:42:01 | examples-with-cassandra/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-harmless-tiger 2>&1 | grep -v "already exists" || true logger.go:42: 07:42:02 | examples-with-cassandra/0-install | service/cassandra created logger.go:42: 07:42:02 | examples-with-cassandra/0-install | statefulset.apps/cassandra created logger.go:42: 07:42:02 | examples-with-cassandra/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:42:05 | examples-with-cassandra/0-install | test step completed 0-install logger.go:42: 07:42:05 | examples-with-cassandra/1-install | starting test step 1-install logger.go:42: 07:42:05 | examples-with-cassandra/1-install | Jaeger:kuttl-test-harmless-tiger/with-cassandra created logger.go:42: 07:42:40 | examples-with-cassandra/1-install | test step completed 1-install logger.go:42: 07:42:40 | examples-with-cassandra/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:42:40 | examples-with-cassandra/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-cassandra /dev/null] logger.go:42: 07:42:41 | examples-with-cassandra/2-smoke-test | Warning: resource jaegers/with-cassandra is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:42:47 | examples-with-cassandra/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:42:48 | examples-with-cassandra/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:42:48 | examples-with-cassandra/2-smoke-test | job.batch/report-span created logger.go:42: 07:42:48 | examples-with-cassandra/2-smoke-test | job.batch/check-span created logger.go:42: 07:43:01 | examples-with-cassandra/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:43:01 | examples-with-cassandra | examples-with-cassandra events from ns kuttl-test-harmless-tiger: logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:02 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-harmless-tiger/cassandra-0 to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:02 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.129.2.30/23] from ovn-kubernetes multus logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:02 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:02 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:02 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:02 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:03 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-harmless-tiger/cassandra-1 to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:03 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.131.0.49/23] from ovn-kubernetes multus logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:03 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:03 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:03 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:03 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:09 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-wnphn Binding Scheduled Successfully assigned kuttl-test-harmless-tiger/with-cassandra-cassandra-schema-job-wnphn to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:09 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-wnphn AddedInterface Add eth0 [10.128.2.50/23] from ovn-kubernetes multus logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:09 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-wnphn.spec.containers{with-cassandra-cassandra-schema-job} Pulling Pulling image "jaegertracing/jaeger-cassandra-schema:1.57.0" kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:09 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job SuccessfulCreate Created pod: with-cassandra-cassandra-schema-job-wnphn job-controller logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:17 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-wnphn.spec.containers{with-cassandra-cassandra-schema-job} Pulled Successfully pulled image "jaegertracing/jaeger-cassandra-schema:1.57.0" in 7.385s (7.385s including waiting). Image size: 381891613 bytes. kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:17 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-wnphn.spec.containers{with-cassandra-cassandra-schema-job} Created Created container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:17 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-wnphn.spec.containers{with-cassandra-cassandra-schema-job} Started Started container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:22 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job Completed Job completed job-controller logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:23 +0000 UTC Normal Pod with-cassandra-5959bcdbd5-s2jjx Binding Scheduled Successfully assigned kuttl-test-harmless-tiger/with-cassandra-5959bcdbd5-s2jjx to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:23 +0000 UTC Normal Pod with-cassandra-5959bcdbd5-s2jjx AddedInterface Add eth0 [10.128.2.51/23] from ovn-kubernetes multus logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:23 +0000 UTC Normal Pod with-cassandra-5959bcdbd5-s2jjx.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:23 +0000 UTC Normal Pod with-cassandra-5959bcdbd5-s2jjx.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:23 +0000 UTC Normal Pod with-cassandra-5959bcdbd5-s2jjx.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:23 +0000 UTC Normal Pod with-cassandra-5959bcdbd5-s2jjx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:23 +0000 UTC Normal Pod with-cassandra-5959bcdbd5-s2jjx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:23 +0000 UTC Normal Pod with-cassandra-5959bcdbd5-s2jjx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:23 +0000 UTC Normal ReplicaSet.apps with-cassandra-5959bcdbd5 SuccessfulCreate Created pod: with-cassandra-5959bcdbd5-s2jjx replicaset-controller logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:23 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-5959bcdbd5 to 1 deployment-controller logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:25 +0000 UTC Warning Pod with-cassandra-5959bcdbd5-s2jjx.spec.containers{jaeger} BackOff Back-off restarting failed container jaeger in pod with-cassandra-5959bcdbd5-s2jjx_kuttl-test-harmless-tiger(75864f6c-54c2-46d4-881f-9b3a8d484184) kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:43 +0000 UTC Normal Pod with-cassandra-54db946cdf-k69pb Binding Scheduled Successfully assigned kuttl-test-harmless-tiger/with-cassandra-54db946cdf-k69pb to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:43 +0000 UTC Normal ReplicaSet.apps with-cassandra-54db946cdf SuccessfulCreate Created pod: with-cassandra-54db946cdf-k69pb replicaset-controller logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:43 +0000 UTC Normal Pod with-cassandra-5959bcdbd5-s2jjx.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:43 +0000 UTC Normal Pod with-cassandra-5959bcdbd5-s2jjx.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:43 +0000 UTC Normal ReplicaSet.apps with-cassandra-5959bcdbd5 SuccessfulDelete Deleted pod: with-cassandra-5959bcdbd5-s2jjx replicaset-controller logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:43 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled down replica set with-cassandra-5959bcdbd5 to 0 from 1 deployment-controller logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:43 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-54db946cdf to 1 deployment-controller logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:44 +0000 UTC Normal Pod with-cassandra-54db946cdf-k69pb AddedInterface Add eth0 [10.128.2.52/23] from ovn-kubernetes multus logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:44 +0000 UTC Normal Pod with-cassandra-54db946cdf-k69pb.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:44 +0000 UTC Normal Pod with-cassandra-54db946cdf-k69pb.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:44 +0000 UTC Normal Pod with-cassandra-54db946cdf-k69pb.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:44 +0000 UTC Normal Pod with-cassandra-54db946cdf-k69pb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:44 +0000 UTC Normal Pod with-cassandra-54db946cdf-k69pb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:44 +0000 UTC Normal Pod with-cassandra-54db946cdf-k69pb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:48 +0000 UTC Normal Pod check-span-pvfd9 Binding Scheduled Successfully assigned kuttl-test-harmless-tiger/check-span-pvfd9 to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:48 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-pvfd9 job-controller logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:48 +0000 UTC Normal Pod report-span-lgg7q Binding Scheduled Successfully assigned kuttl-test-harmless-tiger/report-span-lgg7q to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:48 +0000 UTC Normal Pod report-span-lgg7q AddedInterface Add eth0 [10.129.2.31/23] from ovn-kubernetes multus logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:48 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-lgg7q job-controller logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:49 +0000 UTC Normal Pod check-span-pvfd9 AddedInterface Add eth0 [10.131.0.50/23] from ovn-kubernetes multus logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:49 +0000 UTC Normal Pod check-span-pvfd9.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:49 +0000 UTC Normal Pod check-span-pvfd9.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 802ms (802ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:49 +0000 UTC Normal Pod check-span-pvfd9.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:49 +0000 UTC Normal Pod check-span-pvfd9.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:49 +0000 UTC Normal Pod report-span-lgg7q.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:49 +0000 UTC Normal Pod report-span-lgg7q.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 864ms (864ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:49 +0000 UTC Normal Pod report-span-lgg7q.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:42:49 +0000 UTC Normal Pod report-span-lgg7q.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:43:01 | examples-with-cassandra | 2025-01-27 07:43:00 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:43:01 | examples-with-cassandra | Deleting namespace: kuttl-test-harmless-tiger === CONT kuttl/harness/examples-with-badger logger.go:42: 07:43:13 | examples-with-badger | Creating namespace: kuttl-test-golden-urchin logger.go:42: 07:43:13 | examples-with-badger/0-install | starting test step 0-install logger.go:42: 07:43:13 | examples-with-badger/0-install | Jaeger:kuttl-test-golden-urchin/with-badger created logger.go:42: 07:43:19 | examples-with-badger/0-install | test step completed 0-install logger.go:42: 07:43:19 | examples-with-badger/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:43:19 | examples-with-badger/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger /dev/null] logger.go:42: 07:43:20 | examples-with-badger/1-smoke-test | Warning: resource jaegers/with-badger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:43:29 | examples-with-badger/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:43:30 | examples-with-badger/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:43:30 | examples-with-badger/1-smoke-test | job.batch/report-span created logger.go:42: 07:43:30 | examples-with-badger/1-smoke-test | job.batch/check-span created logger.go:42: 07:43:43 | examples-with-badger/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:43:43 | examples-with-badger | examples-with-badger events from ns kuttl-test-golden-urchin: logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:17 +0000 UTC Normal Pod with-badger-599fb5f747-vv5rf Binding Scheduled Successfully assigned kuttl-test-golden-urchin/with-badger-599fb5f747-vv5rf to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:17 +0000 UTC Normal ReplicaSet.apps with-badger-599fb5f747 SuccessfulCreate Created pod: with-badger-599fb5f747-vv5rf replicaset-controller logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:17 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-599fb5f747 to 1 deployment-controller logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:18 +0000 UTC Normal Pod with-badger-599fb5f747-vv5rf AddedInterface Add eth0 [10.128.2.53/23] from ovn-kubernetes multus logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:18 +0000 UTC Normal Pod with-badger-599fb5f747-vv5rf.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:18 +0000 UTC Normal Pod with-badger-599fb5f747-vv5rf.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:18 +0000 UTC Normal Pod with-badger-599fb5f747-vv5rf.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:18 +0000 UTC Normal Pod with-badger-599fb5f747-vv5rf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:18 +0000 UTC Normal Pod with-badger-599fb5f747-vv5rf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:18 +0000 UTC Normal Pod with-badger-599fb5f747-vv5rf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:26 +0000 UTC Normal Pod with-badger-599fb5f747-vv5rf.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:26 +0000 UTC Normal Pod with-badger-599fb5f747-vv5rf.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:26 +0000 UTC Normal ReplicaSet.apps with-badger-599fb5f747 SuccessfulDelete Deleted pod: with-badger-599fb5f747-vv5rf replicaset-controller logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:26 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled down replica set with-badger-599fb5f747 to 0 from 1 deployment-controller logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:27 +0000 UTC Normal Pod with-badger-574c96575f-fgdtz Binding Scheduled Successfully assigned kuttl-test-golden-urchin/with-badger-574c96575f-fgdtz to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:27 +0000 UTC Normal ReplicaSet.apps with-badger-574c96575f SuccessfulCreate Created pod: with-badger-574c96575f-fgdtz replicaset-controller logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:27 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-574c96575f to 1 deployment-controller logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:28 +0000 UTC Normal Pod with-badger-574c96575f-fgdtz AddedInterface Add eth0 [10.128.2.54/23] from ovn-kubernetes multus logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:28 +0000 UTC Normal Pod with-badger-574c96575f-fgdtz.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:28 +0000 UTC Normal Pod with-badger-574c96575f-fgdtz.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:28 +0000 UTC Normal Pod with-badger-574c96575f-fgdtz.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:28 +0000 UTC Normal Pod with-badger-574c96575f-fgdtz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:28 +0000 UTC Normal Pod with-badger-574c96575f-fgdtz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:28 +0000 UTC Normal Pod with-badger-574c96575f-fgdtz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:30 +0000 UTC Normal Pod check-span-th5wb Binding Scheduled Successfully assigned kuttl-test-golden-urchin/check-span-th5wb to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:30 +0000 UTC Normal Pod check-span-th5wb AddedInterface Add eth0 [10.131.0.51/23] from ovn-kubernetes multus logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:30 +0000 UTC Normal Pod check-span-th5wb.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:30 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-th5wb job-controller logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:30 +0000 UTC Normal Pod report-span-xrpbv Binding Scheduled Successfully assigned kuttl-test-golden-urchin/report-span-xrpbv to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:30 +0000 UTC Normal Pod report-span-xrpbv AddedInterface Add eth0 [10.129.2.32/23] from ovn-kubernetes multus logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:30 +0000 UTC Normal Pod report-span-xrpbv.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:30 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-xrpbv job-controller logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:31 +0000 UTC Normal Pod report-span-xrpbv.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 827ms (827ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:31 +0000 UTC Normal Pod report-span-xrpbv.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:31 +0000 UTC Normal Pod report-span-xrpbv.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:32 +0000 UTC Normal Pod check-span-th5wb.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.569s (1.569s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:32 +0000 UTC Normal Pod check-span-th5wb.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:32 +0000 UTC Normal Pod check-span-th5wb.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:43:43 | examples-with-badger | 2025-01-27 07:43:43 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:43:43 | examples-with-badger | Deleting namespace: kuttl-test-golden-urchin === CONT kuttl/harness/examples-simplest logger.go:42: 07:43:49 | examples-simplest | Creating namespace: kuttl-test-model-insect logger.go:42: 07:43:49 | examples-simplest/0-install | starting test step 0-install logger.go:42: 07:43:49 | examples-simplest/0-install | Jaeger:kuttl-test-model-insect/simplest created logger.go:42: 07:43:57 | examples-simplest/0-install | test step completed 0-install logger.go:42: 07:43:57 | examples-simplest/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:43:57 | examples-simplest/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:43:58 | examples-simplest/1-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:44:04 | examples-simplest/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:44:04 | examples-simplest/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:44:04 | examples-simplest/1-smoke-test | job.batch/report-span created logger.go:42: 07:44:04 | examples-simplest/1-smoke-test | job.batch/check-span created logger.go:42: 07:44:19 | examples-simplest/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:44:19 | examples-simplest | examples-simplest events from ns kuttl-test-model-insect: logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:43:53 +0000 UTC Normal Pod simplest-768687d678-8qpsc Binding Scheduled Successfully assigned kuttl-test-model-insect/simplest-768687d678-8qpsc to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:43:53 +0000 UTC Normal ReplicaSet.apps simplest-768687d678 SuccessfulCreate Created pod: simplest-768687d678-8qpsc replicaset-controller logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:43:53 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-768687d678 to 1 deployment-controller logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:43:54 +0000 UTC Warning Pod simplest-768687d678-8qpsc FailedMount MountVolume.SetUp failed for volume "simplest-collector-tls-config-volume" : secret "simplest-collector-headless-tls" not found kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:43:55 +0000 UTC Normal Pod simplest-768687d678-8qpsc AddedInterface Add eth0 [10.128.2.55/23] from ovn-kubernetes multus logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:43:55 +0000 UTC Normal Pod simplest-768687d678-8qpsc.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:43:55 +0000 UTC Normal Pod simplest-768687d678-8qpsc.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:43:55 +0000 UTC Normal Pod simplest-768687d678-8qpsc.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:43:55 +0000 UTC Normal Pod simplest-768687d678-8qpsc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:43:55 +0000 UTC Normal Pod simplest-768687d678-8qpsc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:43:55 +0000 UTC Normal Pod simplest-768687d678-8qpsc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:00 +0000 UTC Normal Pod simplest-768687d678-8qpsc.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:00 +0000 UTC Normal Pod simplest-768687d678-8qpsc.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:00 +0000 UTC Normal ReplicaSet.apps simplest-768687d678 SuccessfulDelete Deleted pod: simplest-768687d678-8qpsc replicaset-controller logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:00 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-768687d678 to 0 from 1 deployment-controller logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:01 +0000 UTC Normal Pod simplest-549c7dcc-hcbfn Binding Scheduled Successfully assigned kuttl-test-model-insect/simplest-549c7dcc-hcbfn to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:01 +0000 UTC Normal ReplicaSet.apps simplest-549c7dcc SuccessfulCreate Created pod: simplest-549c7dcc-hcbfn replicaset-controller logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:01 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-549c7dcc to 1 deployment-controller logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:02 +0000 UTC Normal Pod simplest-549c7dcc-hcbfn AddedInterface Add eth0 [10.128.2.56/23] from ovn-kubernetes multus logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:02 +0000 UTC Normal Pod simplest-549c7dcc-hcbfn.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:02 +0000 UTC Normal Pod simplest-549c7dcc-hcbfn.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:02 +0000 UTC Normal Pod simplest-549c7dcc-hcbfn.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:02 +0000 UTC Normal Pod simplest-549c7dcc-hcbfn.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:02 +0000 UTC Normal Pod simplest-549c7dcc-hcbfn.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:02 +0000 UTC Normal Pod simplest-549c7dcc-hcbfn.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:04 +0000 UTC Normal Pod check-span-pvjx4 Binding Scheduled Successfully assigned kuttl-test-model-insect/check-span-pvjx4 to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:04 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-pvjx4 job-controller logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:04 +0000 UTC Normal Pod report-span-r5d8g Binding Scheduled Successfully assigned kuttl-test-model-insect/report-span-r5d8g to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:04 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-r5d8g job-controller logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:05 +0000 UTC Normal Pod check-span-pvjx4 AddedInterface Add eth0 [10.131.0.52/23] from ovn-kubernetes multus logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:05 +0000 UTC Normal Pod check-span-pvjx4.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:05 +0000 UTC Normal Pod report-span-r5d8g AddedInterface Add eth0 [10.129.2.33/23] from ovn-kubernetes multus logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:05 +0000 UTC Normal Pod report-span-r5d8g.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:07 +0000 UTC Normal Pod check-span-pvjx4.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.304s (2.304s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:07 +0000 UTC Normal Pod check-span-pvjx4.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:07 +0000 UTC Normal Pod check-span-pvjx4.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:07 +0000 UTC Normal Pod report-span-r5d8g.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.34s (2.34s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:07 +0000 UTC Normal Pod report-span-r5d8g.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:07 +0000 UTC Normal Pod report-span-r5d8g.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:44:19 | examples-simplest | 2025-01-27 07:44:18 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:44:19 | examples-simplest | Deleting namespace: kuttl-test-model-insect === CONT kuttl/harness/examples-simple-prod-with-volumes logger.go:42: 07:44:30 | examples-simple-prod-with-volumes | Ignoring 03-check-volume.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:44:30 | examples-simple-prod-with-volumes | Creating namespace: kuttl-test-epic-badger logger.go:42: 07:44:30 | examples-simple-prod-with-volumes/1-install | starting test step 1-install logger.go:42: 07:44:30 | examples-simple-prod-with-volumes/1-install | Jaeger:kuttl-test-epic-badger/simple-prod created logger.go:42: 07:45:06 | examples-simple-prod-with-volumes/1-install | test step completed 1-install logger.go:42: 07:45:06 | examples-simple-prod-with-volumes/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:45:06 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:45:08 | examples-simple-prod-with-volumes/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:45:14 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:45:14 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:45:14 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/report-span created logger.go:42: 07:45:14 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/check-span created logger.go:42: 07:45:27 | examples-simple-prod-with-volumes/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:45:27 | examples-simple-prod-with-volumes/3-check-volume | starting test step 3-check-volume logger.go:42: 07:45:27 | examples-simple-prod-with-volumes/3-check-volume | running command: [sh -c kubectl exec $(kubectl get pods -n $NAMESPACE -l app=jaeger -l app.kubernetes.io/component=collector -o yaml | /tmp/jaeger-tests/bin/yq e '.items[0].metadata.name') -n $NAMESPACE -- ls /usr/share/elasticsearch/data] logger.go:42: 07:45:28 | examples-simple-prod-with-volumes/3-check-volume | test step completed 3-check-volume logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | examples-simple-prod-with-volumes events from ns kuttl-test-epic-badger: logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:44:36 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestepicbadgersimpleprod-1-84cb7fb9f7 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestepicbadgersimpleprod-1-84cb7fb9jz4j6 replicaset-controller logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:44:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestepicbadgersimpleprod-1-84cb7fb9jz4j6 Binding Scheduled Successfully assigned kuttl-test-epic-badger/elasticsearch-cdm-kuttltestepicbadgersimpleprod-1-84cb7fb9jz4j6 to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:44:36 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestepicbadgersimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestepicbadgersimpleprod-1-84cb7fb9f7 to 1 deployment-controller logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:44:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestepicbadgersimpleprod-1-84cb7fb9jz4j6 AddedInterface Add eth0 [10.131.0.53/23] from ovn-kubernetes multus logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:44:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestepicbadgersimpleprod-1-84cb7fb9jz4j6.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:44:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestepicbadgersimpleprod-1-84cb7fb9jz4j6.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:44:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestepicbadgersimpleprod-1-84cb7fb9jz4j6.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:44:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestepicbadgersimpleprod-1-84cb7fb9jz4j6.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:44:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestepicbadgersimpleprod-1-84cb7fb9jz4j6.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:44:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestepicbadgersimpleprod-1-84cb7fb9jz4j6.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:44:52 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestepicbadgersimpleprod-1-84cb7fb9jz4j6.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:03 +0000 UTC Normal Pod simple-prod-collector-789cb6754c-6l7rb Binding Scheduled Successfully assigned kuttl-test-epic-badger/simple-prod-collector-789cb6754c-6l7rb to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:03 +0000 UTC Warning Pod simple-prod-collector-789cb6754c-6l7rb FailedMount MountVolume.SetUp failed for volume "simple-prod-collector-tls-config-volume" : secret "simple-prod-collector-headless-tls" not found kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:03 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-789cb6754c SuccessfulCreate Created pod: simple-prod-collector-789cb6754c-6l7rb replicaset-controller logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:03 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-789cb6754c to 1 deployment-controller logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:03 +0000 UTC Normal Pod simple-prod-query-676fcdd9f-f7r62 Binding Scheduled Successfully assigned kuttl-test-epic-badger/simple-prod-query-676fcdd9f-f7r62 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:03 +0000 UTC Normal ReplicaSet.apps simple-prod-query-676fcdd9f SuccessfulCreate Created pod: simple-prod-query-676fcdd9f-f7r62 replicaset-controller logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:03 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-676fcdd9f to 1 deployment-controller logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:04 +0000 UTC Normal Pod simple-prod-collector-789cb6754c-6l7rb AddedInterface Add eth0 [10.129.2.34/23] from ovn-kubernetes multus logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:04 +0000 UTC Normal Pod simple-prod-collector-789cb6754c-6l7rb.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:04 +0000 UTC Normal Pod simple-prod-collector-789cb6754c-6l7rb.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:04 +0000 UTC Normal Pod simple-prod-collector-789cb6754c-6l7rb.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:04 +0000 UTC Normal Pod simple-prod-query-676fcdd9f-f7r62 AddedInterface Add eth0 [10.128.2.57/23] from ovn-kubernetes multus logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:04 +0000 UTC Normal Pod simple-prod-query-676fcdd9f-f7r62.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:04 +0000 UTC Normal Pod simple-prod-query-676fcdd9f-f7r62.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:04 +0000 UTC Normal Pod simple-prod-query-676fcdd9f-f7r62.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:04 +0000 UTC Normal Pod simple-prod-query-676fcdd9f-f7r62.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:04 +0000 UTC Normal Pod simple-prod-query-676fcdd9f-f7r62.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:04 +0000 UTC Normal Pod simple-prod-query-676fcdd9f-f7r62.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:04 +0000 UTC Normal Pod simple-prod-query-676fcdd9f-f7r62.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:04 +0000 UTC Normal Pod simple-prod-query-676fcdd9f-f7r62.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:04 +0000 UTC Normal Pod simple-prod-query-676fcdd9f-f7r62.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:10 +0000 UTC Normal Pod simple-prod-query-676fcdd9f-f7r62.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:10 +0000 UTC Normal Pod simple-prod-query-676fcdd9f-f7r62.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:10 +0000 UTC Normal Pod simple-prod-query-676fcdd9f-f7r62.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:10 +0000 UTC Normal ReplicaSet.apps simple-prod-query-676fcdd9f SuccessfulDelete Deleted pod: simple-prod-query-676fcdd9f-f7r62 replicaset-controller logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:10 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-676fcdd9f to 0 from 1 deployment-controller logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:11 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-5b9bc8cd6b to 1 deployment-controller logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:12 +0000 UTC Normal Pod simple-prod-query-5b9bc8cd6b-qcz8x Binding Scheduled Successfully assigned kuttl-test-epic-badger/simple-prod-query-5b9bc8cd6b-qcz8x to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:12 +0000 UTC Normal Pod simple-prod-query-5b9bc8cd6b-qcz8x AddedInterface Add eth0 [10.128.2.58/23] from ovn-kubernetes multus logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:12 +0000 UTC Normal Pod simple-prod-query-5b9bc8cd6b-qcz8x.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:12 +0000 UTC Normal Pod simple-prod-query-5b9bc8cd6b-qcz8x.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:12 +0000 UTC Normal Pod simple-prod-query-5b9bc8cd6b-qcz8x.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:12 +0000 UTC Normal Pod simple-prod-query-5b9bc8cd6b-qcz8x.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:12 +0000 UTC Normal Pod simple-prod-query-5b9bc8cd6b-qcz8x.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:12 +0000 UTC Normal Pod simple-prod-query-5b9bc8cd6b-qcz8x.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:12 +0000 UTC Normal Pod simple-prod-query-5b9bc8cd6b-qcz8x.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:12 +0000 UTC Normal Pod simple-prod-query-5b9bc8cd6b-qcz8x.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:12 +0000 UTC Normal Pod simple-prod-query-5b9bc8cd6b-qcz8x.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:12 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5b9bc8cd6b SuccessfulCreate Created pod: simple-prod-query-5b9bc8cd6b-qcz8x replicaset-controller logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:14 +0000 UTC Normal Pod check-span-hsbpv Binding Scheduled Successfully assigned kuttl-test-epic-badger/check-span-hsbpv to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:14 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-hsbpv job-controller logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:14 +0000 UTC Normal Pod report-span-zdcwg Binding Scheduled Successfully assigned kuttl-test-epic-badger/report-span-zdcwg to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:14 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-zdcwg job-controller logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:15 +0000 UTC Normal Pod check-span-hsbpv AddedInterface Add eth0 [10.129.2.36/23] from ovn-kubernetes multus logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:15 +0000 UTC Normal Pod check-span-hsbpv.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:15 +0000 UTC Normal Pod report-span-zdcwg AddedInterface Add eth0 [10.129.2.35/23] from ovn-kubernetes multus logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:15 +0000 UTC Normal Pod report-span-zdcwg.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:16 +0000 UTC Normal Pod check-span-hsbpv.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.095s (1.095s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:16 +0000 UTC Normal Pod check-span-hsbpv.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:16 +0000 UTC Normal Pod check-span-hsbpv.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:16 +0000 UTC Normal Pod report-span-zdcwg.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.44s (1.44s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:16 +0000 UTC Normal Pod report-span-zdcwg.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:16 +0000 UTC Normal Pod report-span-zdcwg.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:21 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:21 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod simple-prod-collector-789cb6754c-6l7rb horizontal-pod-autoscaler logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:21 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | 2025-01-27 07:45:27 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:45:28 | examples-simple-prod-with-volumes | Deleting namespace: kuttl-test-epic-badger === CONT kuttl/harness/examples-business-application-injected-sidecar logger.go:42: 07:45:40 | examples-business-application-injected-sidecar | Creating namespace: kuttl-test-polished-cub logger.go:42: 07:45:40 | examples-business-application-injected-sidecar/0-install | starting test step 0-install logger.go:42: 07:45:40 | examples-business-application-injected-sidecar/0-install | Deployment:kuttl-test-polished-cub/myapp created logger.go:42: 07:45:40 | examples-business-application-injected-sidecar/0-install | test step completed 0-install logger.go:42: 07:45:40 | examples-business-application-injected-sidecar/1-install | starting test step 1-install logger.go:42: 07:45:40 | examples-business-application-injected-sidecar/1-install | Jaeger:kuttl-test-polished-cub/simplest created logger.go:42: 07:45:51 | examples-business-application-injected-sidecar/1-install | test step completed 1-install logger.go:42: 07:45:51 | examples-business-application-injected-sidecar/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:45:51 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:45:52 | examples-business-application-injected-sidecar/2-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:45:59 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:45:59 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:45:59 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/report-span created logger.go:42: 07:45:59 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/check-span created logger.go:42: 07:46:14 | examples-business-application-injected-sidecar/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | examples-business-application-injected-sidecar events from ns kuttl-test-polished-cub: logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:40 +0000 UTC Normal Pod myapp-fc54f78d8-jcz7m Binding Scheduled Successfully assigned kuttl-test-polished-cub/myapp-fc54f78d8-jcz7m to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:40 +0000 UTC Normal ReplicaSet.apps myapp-fc54f78d8 SuccessfulCreate Created pod: myapp-fc54f78d8-jcz7m replicaset-controller logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:40 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-fc54f78d8 to 1 deployment-controller logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:41 +0000 UTC Normal Pod myapp-fc54f78d8-jcz7m AddedInterface Add eth0 [10.129.2.37/23] from ovn-kubernetes multus logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:41 +0000 UTC Normal Pod myapp-fc54f78d8-jcz7m.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:44 +0000 UTC Normal Pod myapp-649456649d-5jb9b Binding Scheduled Successfully assigned kuttl-test-polished-cub/myapp-649456649d-5jb9b to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:44 +0000 UTC Warning Pod myapp-649456649d-5jb9b FailedMount MountVolume.SetUp failed for volume "simplest-service-ca" : configmap "simplest-service-ca" not found kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:44 +0000 UTC Warning Pod myapp-649456649d-5jb9b FailedMount MountVolume.SetUp failed for volume "simplest-trusted-ca" : configmap "simplest-trusted-ca" not found kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:44 +0000 UTC Normal ReplicaSet.apps myapp-649456649d SuccessfulCreate Created pod: myapp-649456649d-5jb9b replicaset-controller logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:44 +0000 UTC Normal Pod myapp-fc54f78d8-jcz7m.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.414s (3.414s including waiting). Image size: 282912835 bytes. kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:44 +0000 UTC Normal Pod myapp-fc54f78d8-jcz7m.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:44 +0000 UTC Normal Pod myapp-fc54f78d8-jcz7m.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:44 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-649456649d to 1 deployment-controller logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:50 +0000 UTC Normal Pod simplest-54c658cb5b-7h44p Binding Scheduled Successfully assigned kuttl-test-polished-cub/simplest-54c658cb5b-7h44p to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:50 +0000 UTC Normal Pod simplest-54c658cb5b-7h44p AddedInterface Add eth0 [10.128.2.59/23] from ovn-kubernetes multus logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:50 +0000 UTC Normal Pod simplest-54c658cb5b-7h44p.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:50 +0000 UTC Normal Pod simplest-54c658cb5b-7h44p.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:50 +0000 UTC Normal Pod simplest-54c658cb5b-7h44p.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:50 +0000 UTC Normal Pod simplest-54c658cb5b-7h44p.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:50 +0000 UTC Normal Pod simplest-54c658cb5b-7h44p.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:50 +0000 UTC Normal Pod simplest-54c658cb5b-7h44p.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:50 +0000 UTC Normal ReplicaSet.apps simplest-54c658cb5b SuccessfulCreate Created pod: simplest-54c658cb5b-7h44p replicaset-controller logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:50 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-54c658cb5b to 1 deployment-controller logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:52 +0000 UTC Normal Pod myapp-649456649d-5jb9b AddedInterface Add eth0 [10.131.0.55/23] from ovn-kubernetes multus logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:52 +0000 UTC Normal Pod myapp-649456649d-5jb9b.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:52 +0000 UTC Warning Pod myapp-fc54f78d8-jcz7m.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.129.2.37:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:55 +0000 UTC Normal Pod myapp-649456649d-5jb9b.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.068s (3.068s including waiting). Image size: 282912835 bytes. kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:55 +0000 UTC Normal Pod myapp-649456649d-5jb9b.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:55 +0000 UTC Normal Pod myapp-649456649d-5jb9b.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:55 +0000 UTC Normal Pod myapp-649456649d-5jb9b.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:55 +0000 UTC Normal Pod myapp-649456649d-5jb9b.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:55 +0000 UTC Normal Pod myapp-649456649d-5jb9b.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:56 +0000 UTC Normal Pod myapp-fc54f78d8-jcz7m.spec.containers{myapp} Killing Stopping container myapp kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:56 +0000 UTC Normal ReplicaSet.apps myapp-fc54f78d8 SuccessfulDelete Deleted pod: myapp-fc54f78d8-jcz7m replicaset-controller logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:56 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled down replica set myapp-fc54f78d8 to 0 from 1 deployment-controller logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:59 +0000 UTC Normal Pod check-span-trfkv Binding Scheduled Successfully assigned kuttl-test-polished-cub/check-span-trfkv to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:59 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-trfkv job-controller logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:59 +0000 UTC Normal Pod report-span-ptj5f Binding Scheduled Successfully assigned kuttl-test-polished-cub/report-span-ptj5f to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:45:59 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-ptj5f job-controller logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:00 +0000 UTC Normal Pod check-span-trfkv AddedInterface Add eth0 [10.129.2.39/23] from ovn-kubernetes multus logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:00 +0000 UTC Normal Pod check-span-trfkv.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:00 +0000 UTC Normal Pod report-span-ptj5f AddedInterface Add eth0 [10.129.2.38/23] from ovn-kubernetes multus logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:00 +0000 UTC Normal Pod report-span-ptj5f.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:00 +0000 UTC Normal Pod simplest-54c658cb5b-7h44p.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:00 +0000 UTC Normal Pod simplest-54c658cb5b-7h44p.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:00 +0000 UTC Normal ReplicaSet.apps simplest-54c658cb5b SuccessfulDelete Deleted pod: simplest-54c658cb5b-7h44p replicaset-controller logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:00 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-54c658cb5b to 0 from 1 deployment-controller logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:01 +0000 UTC Normal Pod report-span-ptj5f.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 684ms (684ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:01 +0000 UTC Normal Pod report-span-ptj5f.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:01 +0000 UTC Normal Pod report-span-ptj5f.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:01 +0000 UTC Normal Pod simplest-779b67749b-rtxtf Binding Scheduled Successfully assigned kuttl-test-polished-cub/simplest-779b67749b-rtxtf to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:01 +0000 UTC Normal Pod simplest-779b67749b-rtxtf AddedInterface Add eth0 [10.128.2.60/23] from ovn-kubernetes multus logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:01 +0000 UTC Normal Pod simplest-779b67749b-rtxtf.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:01 +0000 UTC Normal Pod simplest-779b67749b-rtxtf.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:01 +0000 UTC Normal Pod simplest-779b67749b-rtxtf.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:01 +0000 UTC Normal Pod simplest-779b67749b-rtxtf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:01 +0000 UTC Normal Pod simplest-779b67749b-rtxtf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:01 +0000 UTC Normal Pod simplest-779b67749b-rtxtf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:01 +0000 UTC Normal ReplicaSet.apps simplest-779b67749b SuccessfulCreate Created pod: simplest-779b67749b-rtxtf replicaset-controller logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:01 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-779b67749b to 1 deployment-controller logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:03 +0000 UTC Warning Pod myapp-649456649d-5jb9b.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.131.0.55:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:04 +0000 UTC Normal Pod check-span-trfkv.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 3.778s (3.778s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:04 +0000 UTC Normal Pod check-span-trfkv.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:04 +0000 UTC Normal Pod check-span-trfkv.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | 2025-01-27 07:46:14 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:46:14 | examples-business-application-injected-sidecar | Deleting namespace: kuttl-test-polished-cub === CONT kuttl/harness/examples-service-types logger.go:42: 07:46:26 | examples-service-types | Creating namespace: kuttl-test-resolved-snail logger.go:42: 07:46:26 | examples-service-types/0-install | starting test step 0-install logger.go:42: 07:46:26 | examples-service-types/0-install | Jaeger:kuttl-test-resolved-snail/service-types created logger.go:42: 07:46:32 | examples-service-types/0-install | test step completed 0-install logger.go:42: 07:46:32 | examples-service-types/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:46:32 | examples-service-types/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE service-types /dev/null] logger.go:42: 07:46:33 | examples-service-types/1-smoke-test | Warning: resource jaegers/service-types is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:46:39 | examples-service-types/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://service-types-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:46:40 | examples-service-types/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:46:40 | examples-service-types/1-smoke-test | job.batch/report-span created logger.go:42: 07:46:40 | examples-service-types/1-smoke-test | job.batch/check-span created logger.go:42: 07:46:53 | examples-service-types/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:46:53 | examples-service-types/2- | starting test step 2- logger.go:42: 07:46:53 | examples-service-types/2- | test step completed 2- logger.go:42: 07:46:53 | examples-service-types | examples-service-types events from ns kuttl-test-resolved-snail: logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:30 +0000 UTC Normal Pod service-types-6fd54746f9-xhfm8 Binding Scheduled Successfully assigned kuttl-test-resolved-snail/service-types-6fd54746f9-xhfm8 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:30 +0000 UTC Normal ReplicaSet.apps service-types-6fd54746f9 SuccessfulCreate Created pod: service-types-6fd54746f9-xhfm8 replicaset-controller logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:30 +0000 UTC Normal Service service-types-collector EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:30 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-6fd54746f9 to 1 deployment-controller logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:31 +0000 UTC Normal Pod service-types-6fd54746f9-xhfm8 AddedInterface Add eth0 [10.128.2.61/23] from ovn-kubernetes multus logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:31 +0000 UTC Normal Pod service-types-6fd54746f9-xhfm8.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:31 +0000 UTC Normal Pod service-types-6fd54746f9-xhfm8.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:31 +0000 UTC Normal Pod service-types-6fd54746f9-xhfm8.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:31 +0000 UTC Normal Pod service-types-6fd54746f9-xhfm8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:31 +0000 UTC Normal Pod service-types-6fd54746f9-xhfm8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:31 +0000 UTC Normal Pod service-types-6fd54746f9-xhfm8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:33 +0000 UTC Normal Service service-types-collector EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:33 +0000 UTC Normal Service service-types-query EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:36 +0000 UTC Normal Pod service-types-6fd54746f9-xhfm8.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:36 +0000 UTC Normal Pod service-types-6fd54746f9-xhfm8.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:36 +0000 UTC Normal ReplicaSet.apps service-types-6fd54746f9 SuccessfulDelete Deleted pod: service-types-6fd54746f9-xhfm8 replicaset-controller logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:36 +0000 UTC Normal Service service-types-query EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:36 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled down replica set service-types-6fd54746f9 to 0 from 1 deployment-controller logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:37 +0000 UTC Normal Pod service-types-886f495bd-bzmlk Binding Scheduled Successfully assigned kuttl-test-resolved-snail/service-types-886f495bd-bzmlk to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:37 +0000 UTC Normal ReplicaSet.apps service-types-886f495bd SuccessfulCreate Created pod: service-types-886f495bd-bzmlk replicaset-controller logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:37 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-886f495bd to 1 deployment-controller logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:38 +0000 UTC Normal Pod service-types-886f495bd-bzmlk AddedInterface Add eth0 [10.128.2.62/23] from ovn-kubernetes multus logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:38 +0000 UTC Normal Pod service-types-886f495bd-bzmlk.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:38 +0000 UTC Normal Pod service-types-886f495bd-bzmlk.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:38 +0000 UTC Normal Pod service-types-886f495bd-bzmlk.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:38 +0000 UTC Normal Pod service-types-886f495bd-bzmlk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:38 +0000 UTC Normal Pod service-types-886f495bd-bzmlk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:38 +0000 UTC Normal Pod service-types-886f495bd-bzmlk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:40 +0000 UTC Normal Pod check-span-v54rf Binding Scheduled Successfully assigned kuttl-test-resolved-snail/check-span-v54rf to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:40 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-v54rf job-controller logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:40 +0000 UTC Normal Pod report-span-qk2sd Binding Scheduled Successfully assigned kuttl-test-resolved-snail/report-span-qk2sd to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:40 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-qk2sd job-controller logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:41 +0000 UTC Normal Pod check-span-v54rf AddedInterface Add eth0 [10.131.0.56/23] from ovn-kubernetes multus logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:41 +0000 UTC Normal Pod check-span-v54rf.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:41 +0000 UTC Normal Pod report-span-qk2sd AddedInterface Add eth0 [10.129.2.40/23] from ovn-kubernetes multus logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:41 +0000 UTC Normal Pod report-span-qk2sd.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:42 +0000 UTC Normal Pod check-span-v54rf.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.132s (1.132s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:42 +0000 UTC Normal Pod check-span-v54rf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:42 +0000 UTC Normal Pod check-span-v54rf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:45 +0000 UTC Normal Pod report-span-qk2sd.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 4.171s (4.171s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:45 +0000 UTC Normal Pod report-span-qk2sd.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:45 +0000 UTC Normal Pod report-span-qk2sd.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:46:53 | examples-service-types | 2025-01-27 07:46:52 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:46:53 | examples-service-types | Deleting namespace: kuttl-test-resolved-snail === CONT kuttl/harness/examples-openshift-with-htpasswd logger.go:42: 07:47:29 | examples-openshift-with-htpasswd | Ignoring 00-install.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:47:29 | examples-openshift-with-htpasswd | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:47:29 | examples-openshift-with-htpasswd | Creating namespace: kuttl-test-neat-moray logger.go:42: 07:47:29 | examples-openshift-with-htpasswd/0-install | starting test step 0-install logger.go:42: 07:47:29 | examples-openshift-with-htpasswd/0-install | Secret:kuttl-test-neat-moray/htpasswd created logger.go:42: 07:47:29 | examples-openshift-with-htpasswd/0-install | test step completed 0-install logger.go:42: 07:47:29 | examples-openshift-with-htpasswd/1-install | starting test step 1-install logger.go:42: 07:47:29 | examples-openshift-with-htpasswd/1-install | Jaeger:kuttl-test-neat-moray/with-htpasswd created logger.go:42: 07:47:35 | examples-openshift-with-htpasswd/1-install | test step completed 1-install logger.go:42: 07:47:35 | examples-openshift-with-htpasswd/2-check-unsecured | starting test step 2-check-unsecured logger.go:42: 07:47:35 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [./ensure-ingress-host.sh] logger.go:42: 07:47:35 | examples-openshift-with-htpasswd/2-check-unsecured | Checking the Ingress host value was populated logger.go:42: 07:47:35 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 0 logger.go:42: 07:47:35 | examples-openshift-with-htpasswd/2-check-unsecured | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 07:47:35 | examples-openshift-with-htpasswd/2-check-unsecured | template was: logger.go:42: 07:47:35 | examples-openshift-with-htpasswd/2-check-unsecured | {.items[0].status.ingress[0].host} logger.go:42: 07:47:35 | examples-openshift-with-htpasswd/2-check-unsecured | object given to jsonpath engine was: logger.go:42: 07:47:35 | examples-openshift-with-htpasswd/2-check-unsecured | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 07:47:35 | examples-openshift-with-htpasswd/2-check-unsecured | logger.go:42: 07:47:35 | examples-openshift-with-htpasswd/2-check-unsecured | logger.go:42: 07:47:45 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 1 logger.go:42: 07:47:45 | examples-openshift-with-htpasswd/2-check-unsecured | Hostname is with-htpasswd-kuttl-test-neat-moray.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:47:45 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:47:45 | examples-openshift-with-htpasswd/2-check-unsecured | Checking an expected HTTP response logger.go:42: 07:47:45 | examples-openshift-with-htpasswd/2-check-unsecured | Running in OpenShift logger.go:42: 07:47:45 | examples-openshift-with-htpasswd/2-check-unsecured | Not using any secret logger.go:42: 07:47:45 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 1/30 the https://with-htpasswd-kuttl-test-neat-moray.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:45 | examples-openshift-with-htpasswd/2-check-unsecured | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:47:45 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 2/30 the https://with-htpasswd-kuttl-test-neat-moray.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:45 | examples-openshift-with-htpasswd/2-check-unsecured | curl response asserted properly logger.go:42: 07:47:45 | examples-openshift-with-htpasswd/2-check-unsecured | test step completed 2-check-unsecured logger.go:42: 07:47:45 | examples-openshift-with-htpasswd/3-check-unauthorized | starting test step 3-check-unauthorized logger.go:42: 07:47:45 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:47:45 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking the Ingress host value was populated logger.go:42: 07:47:45 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 0 logger.go:42: 07:47:45 | examples-openshift-with-htpasswd/3-check-unauthorized | Hostname is with-htpasswd-kuttl-test-neat-moray.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:47:45 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [sh -c JAEGER_USERNAME=wronguser JAEGER_PASSWORD=wrongpassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking an expected HTTP response logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/3-check-unauthorized | Running in OpenShift logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/3-check-unauthorized | Using Jaeger basic authentication logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 1/30 the https://with-htpasswd-kuttl-test-neat-moray.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/3-check-unauthorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 2/30 the https://with-htpasswd-kuttl-test-neat-moray.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/3-check-unauthorized | curl response asserted properly logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/3-check-unauthorized | test step completed 3-check-unauthorized logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/4-check-authorized | starting test step 4-check-authorized logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/4-check-authorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/4-check-authorized | Checking the Ingress host value was populated logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/4-check-authorized | Try number 0 logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/4-check-authorized | Hostname is with-htpasswd-kuttl-test-neat-moray.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/4-check-authorized | running command: [sh -c JAEGER_USERNAME=awesomeuser JAEGER_PASSWORD=awesomepassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE with-htpasswd] logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/4-check-authorized | Checking an expected HTTP response logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/4-check-authorized | Running in OpenShift logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/4-check-authorized | Using Jaeger basic authentication logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/4-check-authorized | Try number 1/30 the https://with-htpasswd-kuttl-test-neat-moray.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/4-check-authorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/4-check-authorized | Try number 2/30 the https://with-htpasswd-kuttl-test-neat-moray.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/4-check-authorized | curl response asserted properly logger.go:42: 07:47:46 | examples-openshift-with-htpasswd/4-check-authorized | test step completed 4-check-authorized logger.go:42: 07:47:46 | examples-openshift-with-htpasswd | examples-openshift-with-htpasswd events from ns kuttl-test-neat-moray: logger.go:42: 07:47:46 | examples-openshift-with-htpasswd | 2025-01-27 07:47:33 +0000 UTC Normal Pod with-htpasswd-867c6965f8-95bpg Binding Scheduled Successfully assigned kuttl-test-neat-moray/with-htpasswd-867c6965f8-95bpg to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:46 | examples-openshift-with-htpasswd | 2025-01-27 07:47:33 +0000 UTC Normal ReplicaSet.apps with-htpasswd-867c6965f8 SuccessfulCreate Created pod: with-htpasswd-867c6965f8-95bpg replicaset-controller logger.go:42: 07:47:46 | examples-openshift-with-htpasswd | 2025-01-27 07:47:33 +0000 UTC Normal Deployment.apps with-htpasswd ScalingReplicaSet Scaled up replica set with-htpasswd-867c6965f8 to 1 deployment-controller logger.go:42: 07:47:46 | examples-openshift-with-htpasswd | 2025-01-27 07:47:34 +0000 UTC Normal Pod with-htpasswd-867c6965f8-95bpg AddedInterface Add eth0 [10.128.2.63/23] from ovn-kubernetes multus logger.go:42: 07:47:46 | examples-openshift-with-htpasswd | 2025-01-27 07:47:34 +0000 UTC Normal Pod with-htpasswd-867c6965f8-95bpg.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:47:46 | examples-openshift-with-htpasswd | 2025-01-27 07:47:34 +0000 UTC Normal Pod with-htpasswd-867c6965f8-95bpg.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:47:46 | examples-openshift-with-htpasswd | 2025-01-27 07:47:34 +0000 UTC Normal Pod with-htpasswd-867c6965f8-95bpg.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:47:46 | examples-openshift-with-htpasswd | 2025-01-27 07:47:34 +0000 UTC Normal Pod with-htpasswd-867c6965f8-95bpg.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:47:46 | examples-openshift-with-htpasswd | 2025-01-27 07:47:34 +0000 UTC Normal Pod with-htpasswd-867c6965f8-95bpg.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:47:46 | examples-openshift-with-htpasswd | 2025-01-27 07:47:34 +0000 UTC Normal Pod with-htpasswd-867c6965f8-95bpg.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:47:46 | examples-openshift-with-htpasswd | Deleting namespace: kuttl-test-neat-moray === CONT kuttl/harness/examples-all-in-one-with-options logger.go:42: 07:47:52 | examples-all-in-one-with-options | Creating namespace: kuttl-test-comic-flounder logger.go:42: 07:47:52 | examples-all-in-one-with-options/0-install | starting test step 0-install logger.go:42: 07:47:52 | examples-all-in-one-with-options/0-install | Jaeger:kuttl-test-comic-flounder/my-jaeger created logger.go:42: 07:47:58 | examples-all-in-one-with-options/0-install | test step completed 0-install logger.go:42: 07:47:58 | examples-all-in-one-with-options/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:47:58 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:47:59 | examples-all-in-one-with-options/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:48:07 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443/jaeger MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:48:07 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:48:08 | examples-all-in-one-with-options/1-smoke-test | job.batch/report-span created logger.go:42: 07:48:08 | examples-all-in-one-with-options/1-smoke-test | job.batch/check-span created logger.go:42: 07:48:20 | examples-all-in-one-with-options/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:48:20 | examples-all-in-one-with-options | examples-all-in-one-with-options events from ns kuttl-test-comic-flounder: logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:47:56 +0000 UTC Normal Pod my-jaeger-6dcfcc9c86-2sksk Binding Scheduled Successfully assigned kuttl-test-comic-flounder/my-jaeger-6dcfcc9c86-2sksk to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:47:56 +0000 UTC Warning Pod my-jaeger-6dcfcc9c86-2sksk FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:47:56 +0000 UTC Warning Pod my-jaeger-6dcfcc9c86-2sksk FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-oauth-proxy-tls" : secret "my-jaeger-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:47:56 +0000 UTC Normal ReplicaSet.apps my-jaeger-6dcfcc9c86 SuccessfulCreate Created pod: my-jaeger-6dcfcc9c86-2sksk replicaset-controller logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:47:56 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6dcfcc9c86 to 1 deployment-controller logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:47:57 +0000 UTC Normal Pod my-jaeger-6dcfcc9c86-2sksk AddedInterface Add eth0 [10.128.2.64/23] from ovn-kubernetes multus logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:47:57 +0000 UTC Normal Pod my-jaeger-6dcfcc9c86-2sksk.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:47:57 +0000 UTC Normal Pod my-jaeger-6dcfcc9c86-2sksk.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:47:57 +0000 UTC Normal Pod my-jaeger-6dcfcc9c86-2sksk.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:47:57 +0000 UTC Normal Pod my-jaeger-6dcfcc9c86-2sksk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:47:57 +0000 UTC Normal Pod my-jaeger-6dcfcc9c86-2sksk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:47:57 +0000 UTC Normal Pod my-jaeger-6dcfcc9c86-2sksk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:04 +0000 UTC Normal Pod my-jaeger-6dcfcc9c86-2sksk.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:04 +0000 UTC Normal Pod my-jaeger-6dcfcc9c86-2sksk.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:04 +0000 UTC Normal ReplicaSet.apps my-jaeger-6dcfcc9c86 SuccessfulDelete Deleted pod: my-jaeger-6dcfcc9c86-2sksk replicaset-controller logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:04 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-6dcfcc9c86 to 0 from 1 deployment-controller logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:05 +0000 UTC Normal Pod my-jaeger-6b4675959d-7vc6b Binding Scheduled Successfully assigned kuttl-test-comic-flounder/my-jaeger-6b4675959d-7vc6b to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:05 +0000 UTC Normal ReplicaSet.apps my-jaeger-6b4675959d SuccessfulCreate Created pod: my-jaeger-6b4675959d-7vc6b replicaset-controller logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:05 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6b4675959d to 1 deployment-controller logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:06 +0000 UTC Normal Pod my-jaeger-6b4675959d-7vc6b AddedInterface Add eth0 [10.128.2.65/23] from ovn-kubernetes multus logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:06 +0000 UTC Normal Pod my-jaeger-6b4675959d-7vc6b.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:06 +0000 UTC Normal Pod my-jaeger-6b4675959d-7vc6b.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:06 +0000 UTC Normal Pod my-jaeger-6b4675959d-7vc6b.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:06 +0000 UTC Normal Pod my-jaeger-6b4675959d-7vc6b.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:06 +0000 UTC Normal Pod my-jaeger-6b4675959d-7vc6b.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:06 +0000 UTC Normal Pod my-jaeger-6b4675959d-7vc6b.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:08 +0000 UTC Normal Pod check-span-gmvhg Binding Scheduled Successfully assigned kuttl-test-comic-flounder/check-span-gmvhg to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:08 +0000 UTC Normal Pod check-span-gmvhg AddedInterface Add eth0 [10.129.2.41/23] from ovn-kubernetes multus logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:08 +0000 UTC Normal Pod check-span-gmvhg.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:08 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-gmvhg job-controller logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:08 +0000 UTC Normal Pod report-span-vszbq Binding Scheduled Successfully assigned kuttl-test-comic-flounder/report-span-vszbq to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:08 +0000 UTC Normal Pod report-span-vszbq AddedInterface Add eth0 [10.131.0.57/23] from ovn-kubernetes multus logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:08 +0000 UTC Normal Pod report-span-vszbq.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:08 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-vszbq job-controller logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:09 +0000 UTC Normal Pod check-span-gmvhg.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 752ms (752ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:09 +0000 UTC Normal Pod check-span-gmvhg.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:09 +0000 UTC Normal Pod check-span-gmvhg.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:09 +0000 UTC Normal Pod report-span-vszbq.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.073s (1.073s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:09 +0000 UTC Normal Pod report-span-vszbq.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:09 +0000 UTC Normal Pod report-span-vszbq.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:48:20 | examples-all-in-one-with-options | 2025-01-27 07:48:20 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:48:20 | examples-all-in-one-with-options | Deleting namespace: kuttl-test-comic-flounder === CONT kuttl/harness/examples-agent-with-priority-class logger.go:42: 07:48:32 | examples-agent-with-priority-class | Creating namespace: kuttl-test-endless-martin logger.go:42: 07:48:32 | examples-agent-with-priority-class/0-install | starting test step 0-install logger.go:42: 07:48:32 | examples-agent-with-priority-class/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 07:48:32 | examples-agent-with-priority-class/0-install | ServiceAccount:kuttl-test-endless-martin/jaeger-agent-daemonset created logger.go:42: 07:48:32 | examples-agent-with-priority-class/0-install | test step completed 0-install logger.go:42: 07:48:32 | examples-agent-with-priority-class/1-add-policy | starting test step 1-add-policy logger.go:42: 07:48:32 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 07:48:32 | examples-agent-with-priority-class/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 07:48:32 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 07:48:37 | examples-agent-with-priority-class/1-add-policy | test step completed 1-add-policy logger.go:42: 07:48:37 | examples-agent-with-priority-class/2-install | starting test step 2-install logger.go:42: 07:48:37 | examples-agent-with-priority-class/2-install | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 07:48:44 | examples-agent-with-priority-class/2-install | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:48:45 | examples-agent-with-priority-class/2-install | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:48:45 | examples-agent-with-priority-class/2-install | job.batch/report-span created logger.go:42: 07:48:45 | examples-agent-with-priority-class/2-install | job.batch/check-span created logger.go:42: 07:48:45 | examples-agent-with-priority-class/2-install | PriorityClass:/high-priority created logger.go:42: 07:48:45 | examples-agent-with-priority-class/2-install | Jaeger:kuttl-test-endless-martin/agent-as-daemonset updated logger.go:42: 07:48:58 | examples-agent-with-priority-class/2-install | test step completed 2-install logger.go:42: 07:48:58 | examples-agent-with-priority-class | examples-agent-with-priority-class events from ns kuttl-test-endless-martin: logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:43 +0000 UTC Normal Pod agent-as-daemonset-5cdf57554-2gsgz Binding Scheduled Successfully assigned kuttl-test-endless-martin/agent-as-daemonset-5cdf57554-2gsgz to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:43 +0000 UTC Normal Pod agent-as-daemonset-5cdf57554-2gsgz AddedInterface Add eth0 [10.128.2.66/23] from ovn-kubernetes multus logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:43 +0000 UTC Normal Pod agent-as-daemonset-5cdf57554-2gsgz.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:43 +0000 UTC Normal Pod agent-as-daemonset-5cdf57554-2gsgz.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:43 +0000 UTC Normal Pod agent-as-daemonset-5cdf57554-2gsgz.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:43 +0000 UTC Normal Pod agent-as-daemonset-5cdf57554-2gsgz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:43 +0000 UTC Normal Pod agent-as-daemonset-5cdf57554-2gsgz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:43 +0000 UTC Normal Pod agent-as-daemonset-5cdf57554-2gsgz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:43 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-5cdf57554 SuccessfulCreate Created pod: agent-as-daemonset-5cdf57554-2gsgz replicaset-controller logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:43 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-5cdf57554 to 1 deployment-controller logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:45 +0000 UTC Normal Pod check-span-gx97t Binding Scheduled Successfully assigned kuttl-test-endless-martin/check-span-gx97t to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:45 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-gx97t job-controller logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:45 +0000 UTC Normal Pod report-span-4wv29 Binding Scheduled Successfully assigned kuttl-test-endless-martin/report-span-4wv29 to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:45 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-4wv29 job-controller logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:46 +0000 UTC Normal Pod check-span-gx97t AddedInterface Add eth0 [10.131.0.58/23] from ovn-kubernetes multus logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:46 +0000 UTC Normal Pod check-span-gx97t.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:46 +0000 UTC Normal Pod report-span-4wv29 AddedInterface Add eth0 [10.129.2.42/23] from ovn-kubernetes multus logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:46 +0000 UTC Normal Pod report-span-4wv29.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:47 +0000 UTC Normal Pod check-span-gx97t.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.651s (1.651s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:47 +0000 UTC Normal Pod check-span-gx97t.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:47 +0000 UTC Normal Pod check-span-gx97t.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:47 +0000 UTC Normal Pod report-span-4wv29.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 945ms (945ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:47 +0000 UTC Normal Pod report-span-4wv29.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:47 +0000 UTC Normal Pod report-span-4wv29.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:49 +0000 UTC Warning DaemonSet.apps agent-as-daemonset-agent-daemonset FailedCreate Error creating: pods "agent-as-daemonset-agent-daemonset-" is forbidden: unable to validate against any security context constraint: [provider "anyuid": Forbidden: not usable by user or serviceaccount, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5775: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5778: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6831: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6832: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 14271: Host ports are not allowed to be used, provider "restricted": Forbidden: not usable by user or serviceaccount, provider "nonroot-v2": Forbidden: not usable by user or serviceaccount, provider "nonroot": Forbidden: not usable by user or serviceaccount, provider "hostmount-anyuid": Forbidden: not usable by user or serviceaccount, provider "elasticsearch-scc": Forbidden: not usable by user or serviceaccount, provider "machine-api-termination-handler": Forbidden: not usable by user or serviceaccount, provider "daemonset-with-hostport": Forbidden: not usable by user or serviceaccount, provider "hostnetwork-v2": Forbidden: not usable by user or serviceaccount, provider "hostnetwork": Forbidden: not usable by user or serviceaccount, provider "hostaccess": Forbidden: not usable by user or serviceaccount, provider "node-exporter": Forbidden: not usable by user or serviceaccount, provider "privileged": Forbidden: not usable by user or serviceaccount] daemonset-controller logger.go:42: 07:48:58 | examples-agent-with-priority-class | 2025-01-27 07:48:58 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:48:58 | examples-agent-with-priority-class | Deleting namespace: kuttl-test-endless-martin === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (608.90s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.08s) --- PASS: kuttl/harness/examples-simple-prod (68.55s) --- PASS: kuttl/harness/examples-with-sampling (109.76s) --- PASS: kuttl/harness/examples-with-cassandra (71.84s) --- PASS: kuttl/harness/examples-with-badger (36.24s) --- PASS: kuttl/harness/examples-simplest (40.59s) --- PASS: kuttl/harness/examples-simple-prod-with-volumes (70.09s) --- PASS: kuttl/harness/examples-business-application-injected-sidecar (45.91s) --- PASS: kuttl/harness/examples-service-types (62.83s) --- PASS: kuttl/harness/examples-openshift-with-htpasswd (23.03s) --- PASS: kuttl/harness/examples-all-in-one-with-options (40.21s) --- PASS: kuttl/harness/examples-agent-with-priority-class (33.71s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name examples --report --output /logs/artifacts/examples.xml ./artifacts/kuttl-report.xml time="2025-01-27T07:49:06Z" level=debug msg="Setting a new name for the test suites" time="2025-01-27T07:49:06Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-01-27T07:49:06Z" level=debug msg="normalizing test case names" time="2025-01-27T07:49:06Z" level=debug msg="examples/artifacts -> examples_artifacts" time="2025-01-27T07:49:06Z" level=debug msg="examples/examples-simple-prod -> examples_examples_simple_prod" time="2025-01-27T07:49:06Z" level=debug msg="examples/examples-with-sampling -> examples_examples_with_sampling" time="2025-01-27T07:49:06Z" level=debug msg="examples/examples-with-cassandra -> examples_examples_with_cassandra" time="2025-01-27T07:49:06Z" level=debug msg="examples/examples-with-badger -> examples_examples_with_badger" time="2025-01-27T07:49:06Z" level=debug msg="examples/examples-simplest -> examples_examples_simplest" time="2025-01-27T07:49:06Z" level=debug msg="examples/examples-simple-prod-with-volumes -> examples_examples_simple_prod_with_volumes" time="2025-01-27T07:49:06Z" level=debug msg="examples/examples-business-application-injected-sidecar -> examples_examples_business_application_injected_sidecar" time="2025-01-27T07:49:06Z" level=debug msg="examples/examples-service-types -> examples_examples_service_types" time="2025-01-27T07:49:06Z" level=debug msg="examples/examples-openshift-with-htpasswd -> examples_examples_openshift_with_htpasswd" time="2025-01-27T07:49:06Z" level=debug msg="examples/examples-all-in-one-with-options -> examples_examples_all_in_one_with_options" time="2025-01-27T07:49:06Z" level=debug msg="examples/examples-agent-with-priority-class -> examples_examples_agent_with_priority_class" +---------------------------------------------------------+--------+ | NAME | RESULT | +---------------------------------------------------------+--------+ | examples_artifacts | passed | | examples_examples_simple_prod | passed | | examples_examples_with_sampling | passed | | examples_examples_with_cassandra | passed | | examples_examples_with_badger | passed | | examples_examples_simplest | passed | | examples_examples_simple_prod_with_volumes | passed | | examples_examples_business_application_injected_sidecar | passed | | examples_examples_service_types | passed | | examples_examples_openshift_with_htpasswd | passed | | examples_examples_all_in_one_with_options | passed | | examples_examples_agent_with_priority_class | passed | +---------------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh generate false true + '[' 3 -ne 3 ']' + test_suite_name=generate + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/generate.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-generate make[2]: Entering directory '/tmp/jaeger-tests' test -s /tmp/jaeger-tests/bin/operator-sdk || curl -sLo /tmp/jaeger-tests/bin/operator-sdk https://github.com/operator-framework/operator-sdk/releases/download/v1.32.0/operator-sdk_`go env GOOS`_`go env GOARCH` ./hack/install/install-golangci-lint.sh Installing golangci-lint golangci-lint 1.55.2 is installed already ./hack/install/install-goimports.sh Installing goimports Try 0... go install golang.org/x/tools/cmd/goimports@v0.1.12 >>>> Formatting code... ./.ci/format.sh >>>> Building... ./hack/install/install-dependencies.sh Installing go dependencies Try 0... go mod download GOOS= GOARCH= CGO_ENABLED=0 GO111MODULE=on go build -ldflags "-X "github.com/jaegertracing/jaeger-operator/pkg/version".version="1.62.0" -X "github.com/jaegertracing/jaeger-operator/pkg/version".buildDate=2025-01-27T07:49:08Z -X "github.com/jaegertracing/jaeger-operator/pkg/version".defaultJaeger="1.62.0"" -o "bin/jaeger-operator" main.go JAEGER_VERSION="1.62.0" ./tests/e2e/generate/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-25-163410 True False 38m Cluster version is 4.18.0-0.nightly-2025-01-25-163410' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-25-163410 True False 38m Cluster version is 4.18.0-0.nightly-2025-01-25-163410' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/generate/render.sh ++ export SUITE_DIR=./tests/e2e/generate ++ SUITE_DIR=./tests/e2e/generate ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/generate ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test generate 'This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 2 -ne 2 ']' + test_name=generate + message='This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/generate/_build + '[' _build '!=' _build ']' + rm -rf generate + warning 'generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed\e[0m' WAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running generate E2E tests' Running generate E2E tests + cd tests/e2e/generate/_build + set +e + KUBECONFIG=/tmp/kubeconfig-4053090600 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:49:23 | artifacts | Creating namespace: kuttl-test-magnetic-airedale logger.go:42: 07:49:23 | artifacts | artifacts events from ns kuttl-test-magnetic-airedale: logger.go:42: 07:49:23 | artifacts | Deleting namespace: kuttl-test-magnetic-airedale === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.09s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.05s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name generate --report --output /logs/artifacts/generate.xml ./artifacts/kuttl-report.xml time="2025-01-27T07:49:29Z" level=debug msg="Setting a new name for the test suites" time="2025-01-27T07:49:29Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-01-27T07:49:29Z" level=debug msg="normalizing test case names" time="2025-01-27T07:49:29Z" level=debug msg="generate/artifacts -> generate_artifacts" +--------------------+--------+ | NAME | RESULT | +--------------------+--------+ | generate_artifacts | passed | +--------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh upgrade false true + '[' 3 -ne 3 ']' + test_suite_name=upgrade + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/upgrade.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-upgrade make[2]: Entering directory '/tmp/jaeger-tests' make docker JAEGER_VERSION=1.62.1 IMG="quay.io//jaeger-operator:next" make[3]: Entering directory '/tmp/jaeger-tests' [ ! -z "true" ] || docker build --build-arg=GOPROXY= --build-arg=VERSION="1.62.0" --build-arg=JAEGER_VERSION=1.62.1 --build-arg=TARGETARCH= --build-arg VERSION_DATE=2025-01-27T07:49:30Z --build-arg VERSION_PKG="github.com/jaegertracing/jaeger-operator/pkg/version" -t "quay.io//jaeger-operator:next" . make[3]: Leaving directory '/tmp/jaeger-tests' touch build-e2e-upgrade-image SKIP_ES_EXTERNAL=true IMG=quay.io//jaeger-operator:"1.62.0" JAEGER_OPERATOR_VERSION="1.62.0" JAEGER_VERSION="1.62.0" ./tests/e2e/upgrade/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-25-163410 True False 39m Cluster version is 4.18.0-0.nightly-2025-01-25-163410' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-25-163410 True False 39m Cluster version is 4.18.0-0.nightly-2025-01-25-163410' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/upgrade/render.sh ++ export SUITE_DIR=./tests/e2e/upgrade ++ SUITE_DIR=./tests/e2e/upgrade ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/upgrade ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + export JAEGER_NAME + '[' true = true ']' + skip_test upgrade 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade + warning 'upgrade: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade: Test not supported in OpenShift\e[0m' WAR: upgrade: Test not supported in OpenShift + '[' true = true ']' + skip_test upgrade-from-latest-release 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade-from-latest-release + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade-from-latest-release + warning 'upgrade-from-latest-release: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade-from-latest-release: Test not supported in OpenShift\e[0m' WAR: upgrade-from-latest-release: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running upgrade E2E tests' Running upgrade E2E tests + cd tests/e2e/upgrade/_build + set +e + KUBECONFIG=/tmp/kubeconfig-4053090600 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:49:31 | artifacts | Creating namespace: kuttl-test-huge-anchovy logger.go:42: 07:49:31 | artifacts | artifacts events from ns kuttl-test-huge-anchovy: logger.go:42: 07:49:31 | artifacts | Deleting namespace: kuttl-test-huge-anchovy === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.08s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.04s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name upgrade --report --output /logs/artifacts/upgrade.xml ./artifacts/kuttl-report.xml time="2025-01-27T07:49:37Z" level=debug msg="Setting a new name for the test suites" time="2025-01-27T07:49:37Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-01-27T07:49:37Z" level=debug msg="normalizing test case names" time="2025-01-27T07:49:37Z" level=debug msg="upgrade/artifacts -> upgrade_artifacts" +-------------------+--------+ | NAME | RESULT | +-------------------+--------+ | upgrade_artifacts | passed | +-------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh sidecar false true + '[' 3 -ne 3 ']' + test_suite_name=sidecar + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/sidecar.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-sidecar make[2]: Entering directory '/tmp/jaeger-tests' ./tests/e2e/sidecar/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-25-163410 True False 39m Cluster version is 4.18.0-0.nightly-2025-01-25-163410' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-25-163410 True False 39m Cluster version is 4.18.0-0.nightly-2025-01-25-163410' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/sidecar/render.sh ++ export SUITE_DIR=./tests/e2e/sidecar ++ SUITE_DIR=./tests/e2e/sidecar ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/sidecar ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + jaeger_service_name=order + start_test sidecar-deployment + '[' 1 -ne 1 ']' + test_name=sidecar-deployment + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-deployment' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-deployment\e[0m' Rendering files for test sidecar-deployment + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build + '[' _build '!=' _build ']' + mkdir -p sidecar-deployment + cd sidecar-deployment + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-namespace + '[' 1 -ne 1 ']' + test_name=sidecar-namespace + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-namespace' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-namespace\e[0m' Rendering files for test sidecar-namespace + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-deployment + '[' sidecar-deployment '!=' _build ']' + cd .. + mkdir -p sidecar-namespace + cd sidecar-namespace + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-skip-webhook + '[' 1 -ne 1 ']' + test_name=sidecar-skip-webhook + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-skip-webhook' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-skip-webhook\e[0m' Rendering files for test sidecar-skip-webhook + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-namespace + '[' sidecar-namespace '!=' _build ']' + cd .. + mkdir -p sidecar-skip-webhook + cd sidecar-skip-webhook + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running sidecar E2E tests' Running sidecar E2E tests + cd tests/e2e/sidecar/_build + set +e + KUBECONFIG=/tmp/kubeconfig-4053090600 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/sidecar-deployment === PAUSE kuttl/harness/sidecar-deployment === RUN kuttl/harness/sidecar-namespace === PAUSE kuttl/harness/sidecar-namespace === RUN kuttl/harness/sidecar-skip-webhook === PAUSE kuttl/harness/sidecar-skip-webhook === CONT kuttl/harness/artifacts logger.go:42: 07:49:45 | artifacts | Creating namespace: kuttl-test-smiling-caiman logger.go:42: 07:49:45 | artifacts | artifacts events from ns kuttl-test-smiling-caiman: logger.go:42: 07:49:45 | artifacts | Deleting namespace: kuttl-test-smiling-caiman === CONT kuttl/harness/sidecar-namespace logger.go:42: 07:49:51 | sidecar-namespace | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:49:51 | sidecar-namespace | Creating namespace: kuttl-test-clean-crappie logger.go:42: 07:49:51 | sidecar-namespace/0-install | starting test step 0-install logger.go:42: 07:49:51 | sidecar-namespace/0-install | Jaeger:kuttl-test-clean-crappie/agent-as-sidecar created logger.go:42: 07:49:58 | sidecar-namespace/0-install | test step completed 0-install logger.go:42: 07:49:58 | sidecar-namespace/1-install | starting test step 1-install logger.go:42: 07:49:58 | sidecar-namespace/1-install | Deployment:kuttl-test-clean-crappie/vertx-create-span-sidecar created logger.go:42: 07:49:59 | sidecar-namespace/1-install | test step completed 1-install logger.go:42: 07:49:59 | sidecar-namespace/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:49:59 | sidecar-namespace/2-enable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="true"] logger.go:42: 07:49:59 | sidecar-namespace/2-enable-injection | namespace/kuttl-test-clean-crappie annotated logger.go:42: 07:50:05 | sidecar-namespace/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:50:05 | sidecar-namespace/3-find-service | starting test step 3-find-service logger.go:42: 07:50:05 | sidecar-namespace/3-find-service | Job:kuttl-test-clean-crappie/00-find-service created logger.go:42: 07:50:17 | sidecar-namespace/3-find-service | test step completed 3-find-service logger.go:42: 07:50:17 | sidecar-namespace/4-other-instance | starting test step 4-other-instance logger.go:42: 07:50:17 | sidecar-namespace/4-other-instance | Jaeger:kuttl-test-clean-crappie/agent-as-sidecar2 created logger.go:42: 07:50:26 | sidecar-namespace/4-other-instance | test step completed 4-other-instance logger.go:42: 07:50:26 | sidecar-namespace/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:50:26 | sidecar-namespace/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:50:26 | sidecar-namespace/6-find-service | starting test step 6-find-service logger.go:42: 07:50:26 | sidecar-namespace/6-find-service | Job:kuttl-test-clean-crappie/01-find-service created logger.go:42: 07:50:47 | sidecar-namespace/6-find-service | test step completed 6-find-service logger.go:42: 07:50:47 | sidecar-namespace/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:50:47 | sidecar-namespace/7-disable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="false"] logger.go:42: 07:50:48 | sidecar-namespace/7-disable-injection | namespace/kuttl-test-clean-crappie annotated logger.go:42: 07:50:50 | sidecar-namespace/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:50:50 | sidecar-namespace | sidecar-namespace events from ns kuttl-test-clean-crappie: logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:49:55 +0000 UTC Normal Pod agent-as-sidecar-65f68585d-477rz Binding Scheduled Successfully assigned kuttl-test-clean-crappie/agent-as-sidecar-65f68585d-477rz to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:49:55 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-65f68585d SuccessfulCreate Created pod: agent-as-sidecar-65f68585d-477rz replicaset-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:49:55 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-65f68585d to 1 deployment-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:49:56 +0000 UTC Normal Pod agent-as-sidecar-65f68585d-477rz AddedInterface Add eth0 [10.129.2.43/23] from ovn-kubernetes multus logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:49:56 +0000 UTC Normal Pod agent-as-sidecar-65f68585d-477rz.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:49:57 +0000 UTC Normal Pod agent-as-sidecar-65f68585d-477rz.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" in 1.616s (1.616s including waiting). Image size: 147411949 bytes. kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:49:57 +0000 UTC Normal Pod agent-as-sidecar-65f68585d-477rz.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:49:57 +0000 UTC Normal Pod agent-as-sidecar-65f68585d-477rz.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:49:58 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-h4cft Binding Scheduled Successfully assigned kuttl-test-clean-crappie/vertx-create-span-sidecar-797645c8fc-h4cft to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:49:58 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulCreate Created pod: vertx-create-span-sidecar-797645c8fc-h4cft replicaset-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:49:58 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-797645c8fc to 1 deployment-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:49:59 +0000 UTC Normal Pod vertx-create-span-sidecar-6847f86c8f-wgh5d Binding Scheduled Successfully assigned kuttl-test-clean-crappie/vertx-create-span-sidecar-6847f86c8f-wgh5d to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:49:59 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6847f86c8f SuccessfulCreate Created pod: vertx-create-span-sidecar-6847f86c8f-wgh5d replicaset-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:49:59 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-h4cft AddedInterface Add eth0 [10.131.0.59/23] from ovn-kubernetes multus logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:49:59 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-h4cft.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:49:59 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-h4cft.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:49:59 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-h4cft.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:49:59 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6847f86c8f to 1 deployment-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:00 +0000 UTC Normal Pod vertx-create-span-sidecar-6847f86c8f-wgh5d AddedInterface Add eth0 [10.128.2.67/23] from ovn-kubernetes multus logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:00 +0000 UTC Normal Pod vertx-create-span-sidecar-6847f86c8f-wgh5d.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:03 +0000 UTC Normal Pod vertx-create-span-sidecar-6847f86c8f-wgh5d.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.518s (3.518s including waiting). Image size: 282912835 bytes. kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:03 +0000 UTC Normal Pod vertx-create-span-sidecar-6847f86c8f-wgh5d.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:03 +0000 UTC Normal Pod vertx-create-span-sidecar-6847f86c8f-wgh5d.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:03 +0000 UTC Normal Pod vertx-create-span-sidecar-6847f86c8f-wgh5d.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:03 +0000 UTC Normal Pod vertx-create-span-sidecar-6847f86c8f-wgh5d.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:03 +0000 UTC Normal Pod vertx-create-span-sidecar-6847f86c8f-wgh5d.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:05 +0000 UTC Normal Pod 00-find-service-gpt7n Binding Scheduled Successfully assigned kuttl-test-clean-crappie/00-find-service-gpt7n to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:05 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-gpt7n job-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:06 +0000 UTC Normal Pod 00-find-service-gpt7n AddedInterface Add eth0 [10.129.2.44/23] from ovn-kubernetes multus logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:06 +0000 UTC Normal Pod 00-find-service-gpt7n.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:07 +0000 UTC Normal Pod 00-find-service-gpt7n.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 916ms (916ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:07 +0000 UTC Normal Pod 00-find-service-gpt7n.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:07 +0000 UTC Normal Pod 00-find-service-gpt7n.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:07 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-h4cft.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.59:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:07 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-h4cft.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.59:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:09 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-h4cft.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:10 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-h4cft.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.59:8080/": read tcp 10.131.0.2:59382->10.131.0.59:8080: read: connection reset by peer kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:10 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-h4cft.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.59:8080/": dial tcp 10.131.0.59:8080: connect: connection refused kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:11 +0000 UTC Warning Pod vertx-create-span-sidecar-6847f86c8f-wgh5d.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.67:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:11 +0000 UTC Warning Pod vertx-create-span-sidecar-6847f86c8f-wgh5d.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.67:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:14 +0000 UTC Normal Pod vertx-create-span-sidecar-6847f86c8f-wgh5d.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:14 +0000 UTC Warning Pod vertx-create-span-sidecar-6847f86c8f-wgh5d.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.67:8080/": read tcp 10.128.2.2:59050->10.128.2.67:8080: read: connection reset by peer kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:14 +0000 UTC Warning Pod vertx-create-span-sidecar-6847f86c8f-wgh5d.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.67:8080/": dial tcp 10.128.2.67:8080: connect: connection refused kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:14 +0000 UTC Normal Pod vertx-create-span-sidecar-6847f86c8f-wgh5d.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:17 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:21 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-h4cft.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.59:8080/": read tcp 10.131.0.2:36482->10.131.0.59:8080: read: connection reset by peer kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:24 +0000 UTC Normal Pod agent-as-sidecar2-56c85b5b68-2r2w8 Binding Scheduled Successfully assigned kuttl-test-clean-crappie/agent-as-sidecar2-56c85b5b68-2r2w8 to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:24 +0000 UTC Normal Pod agent-as-sidecar2-56c85b5b68-2r2w8 AddedInterface Add eth0 [10.129.2.45/23] from ovn-kubernetes multus logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:24 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-56c85b5b68 SuccessfulCreate Created pod: agent-as-sidecar2-56c85b5b68-2r2w8 replicaset-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:24 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-56c85b5b68 to 1 deployment-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:25 +0000 UTC Normal Pod agent-as-sidecar2-56c85b5b68-2r2w8.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:25 +0000 UTC Normal Pod agent-as-sidecar2-56c85b5b68-2r2w8.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:25 +0000 UTC Normal Pod agent-as-sidecar2-56c85b5b68-2r2w8.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:26 +0000 UTC Normal Pod 01-find-service-6xx7n Binding Scheduled Successfully assigned kuttl-test-clean-crappie/01-find-service-6xx7n to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:26 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-6xx7n job-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:26 +0000 UTC Normal Pod agent-as-sidecar-65f68585d-477rz.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:27 +0000 UTC Normal Pod 01-find-service-6xx7n AddedInterface Add eth0 [10.131.0.60/23] from ovn-kubernetes multus logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:27 +0000 UTC Normal Pod 01-find-service-6xx7n.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:28 +0000 UTC Normal Pod 01-find-service-6xx7n.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.207s (1.207s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:28 +0000 UTC Normal Pod 01-find-service-6xx7n.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:28 +0000 UTC Normal Pod 01-find-service-6xx7n.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:30 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulDelete Deleted pod: vertx-create-span-sidecar-797645c8fc-h4cft replicaset-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:30 +0000 UTC Normal Pod vertx-create-span-sidecar-855dcc9455-5fmf8 Binding Scheduled Successfully assigned kuttl-test-clean-crappie/vertx-create-span-sidecar-855dcc9455-5fmf8 to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:30 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-855dcc9455 SuccessfulCreate Created pod: vertx-create-span-sidecar-855dcc9455-5fmf8 replicaset-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:30 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-797645c8fc to 0 from 1 deployment-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:30 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-855dcc9455 to 1 from 0 deployment-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:31 +0000 UTC Normal Pod vertx-create-span-sidecar-855dcc9455-5fmf8 AddedInterface Add eth0 [10.129.2.46/23] from ovn-kubernetes multus logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:31 +0000 UTC Normal Pod vertx-create-span-sidecar-855dcc9455-5fmf8.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:31 +0000 UTC Normal Pod vertx-create-span-sidecar-855dcc9455-5fmf8.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:31 +0000 UTC Normal Pod vertx-create-span-sidecar-855dcc9455-5fmf8.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:31 +0000 UTC Normal Pod vertx-create-span-sidecar-855dcc9455-5fmf8.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:33 +0000 UTC Normal Pod vertx-create-span-sidecar-855dcc9455-5fmf8.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" in 2.347s (2.347s including waiting). Image size: 115697219 bytes. kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:33 +0000 UTC Normal Pod vertx-create-span-sidecar-855dcc9455-5fmf8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:33 +0000 UTC Normal Pod vertx-create-span-sidecar-855dcc9455-5fmf8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:39 +0000 UTC Warning Pod vertx-create-span-sidecar-855dcc9455-5fmf8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.46:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:39 +0000 UTC Warning Pod vertx-create-span-sidecar-855dcc9455-5fmf8.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.46:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:41 +0000 UTC Normal Pod vertx-create-span-sidecar-855dcc9455-5fmf8.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:42 +0000 UTC Warning Pod vertx-create-span-sidecar-855dcc9455-5fmf8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.46:8080/": read tcp 10.129.2.2:56868->10.129.2.46:8080: read: connection reset by peer kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:42 +0000 UTC Warning Pod vertx-create-span-sidecar-855dcc9455-5fmf8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.46:8080/": dial tcp 10.129.2.46:8080: connect: connection refused kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:47 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:48 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6847f86c8f SuccessfulDelete Deleted pod: vertx-create-span-sidecar-6847f86c8f-wgh5d replicaset-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:48 +0000 UTC Normal Pod vertx-create-span-sidecar-8565fb66c6-wkrrb Binding Scheduled Successfully assigned kuttl-test-clean-crappie/vertx-create-span-sidecar-8565fb66c6-wkrrb to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:48 +0000 UTC Normal Pod vertx-create-span-sidecar-8565fb66c6-wkrrb AddedInterface Add eth0 [10.131.0.61/23] from ovn-kubernetes multus logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:48 +0000 UTC Normal Pod vertx-create-span-sidecar-8565fb66c6-wkrrb.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:48 +0000 UTC Normal Pod vertx-create-span-sidecar-8565fb66c6-wkrrb.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:48 +0000 UTC Normal Pod vertx-create-span-sidecar-8565fb66c6-wkrrb.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:48 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-8565fb66c6 SuccessfulCreate Created pod: vertx-create-span-sidecar-8565fb66c6-wkrrb replicaset-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:48 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-6847f86c8f to 0 from 1 deployment-controller logger.go:42: 07:50:50 | sidecar-namespace | 2025-01-27 07:50:48 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-8565fb66c6 to 1 from 0 deployment-controller logger.go:42: 07:50:50 | sidecar-namespace | Deleting namespace: kuttl-test-clean-crappie === CONT kuttl/harness/sidecar-skip-webhook logger.go:42: 07:50:56 | sidecar-skip-webhook | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:50:56 | sidecar-skip-webhook | Creating namespace: kuttl-test-top-sawfly logger.go:42: 07:50:56 | sidecar-skip-webhook/0-install | starting test step 0-install logger.go:42: 07:50:56 | sidecar-skip-webhook/0-install | Jaeger:kuttl-test-top-sawfly/agent-as-sidecar created logger.go:42: 07:51:04 | sidecar-skip-webhook/0-install | test step completed 0-install logger.go:42: 07:51:04 | sidecar-skip-webhook/1-install | starting test step 1-install logger.go:42: 07:51:04 | sidecar-skip-webhook/1-install | Deployment:kuttl-test-top-sawfly/vertx-create-span-sidecar created logger.go:42: 07:51:05 | sidecar-skip-webhook/1-install | test step completed 1-install logger.go:42: 07:51:05 | sidecar-skip-webhook/2-add-anotation-and-label | starting test step 2-add-anotation-and-label logger.go:42: 07:51:05 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name=jaeger-operator --namespace kuttl-test-top-sawfly] logger.go:42: 07:51:05 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar labeled logger.go:42: 07:51:05 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-top-sawfly] logger.go:42: 07:51:05 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:51:05 | sidecar-skip-webhook/2-add-anotation-and-label | test step completed 2-add-anotation-and-label logger.go:42: 07:51:05 | sidecar-skip-webhook/3-remove-label | starting test step 3-remove-label logger.go:42: 07:51:05 | sidecar-skip-webhook/3-remove-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name- --namespace kuttl-test-top-sawfly] logger.go:42: 07:51:05 | sidecar-skip-webhook/3-remove-label | deployment.apps/vertx-create-span-sidecar unlabeled logger.go:42: 07:51:07 | sidecar-skip-webhook/3-remove-label | test step completed 3-remove-label logger.go:42: 07:51:07 | sidecar-skip-webhook | sidecar-skip-webhook events from ns kuttl-test-top-sawfly: logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:00 +0000 UTC Normal Pod agent-as-sidecar-768d7bfdb6-gxmhs Binding Scheduled Successfully assigned kuttl-test-top-sawfly/agent-as-sidecar-768d7bfdb6-gxmhs to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:00 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-768d7bfdb6 SuccessfulCreate Created pod: agent-as-sidecar-768d7bfdb6-gxmhs replicaset-controller logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:00 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-768d7bfdb6 to 1 deployment-controller logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:02 +0000 UTC Normal Pod agent-as-sidecar-768d7bfdb6-gxmhs AddedInterface Add eth0 [10.129.2.47/23] from ovn-kubernetes multus logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:02 +0000 UTC Normal Pod agent-as-sidecar-768d7bfdb6-gxmhs.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:02 +0000 UTC Normal Pod agent-as-sidecar-768d7bfdb6-gxmhs.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:02 +0000 UTC Normal Pod agent-as-sidecar-768d7bfdb6-gxmhs.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:04 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-kt4ws Binding Scheduled Successfully assigned kuttl-test-top-sawfly/vertx-create-span-sidecar-797645c8fc-kt4ws to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:04 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-kt4ws AddedInterface Add eth0 [10.128.2.68/23] from ovn-kubernetes multus logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:04 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-kt4ws.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:04 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-kt4ws.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:04 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-kt4ws.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:04 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulCreate Created pod: vertx-create-span-sidecar-797645c8fc-kt4ws replicaset-controller logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:04 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-797645c8fc to 1 deployment-controller logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:05 +0000 UTC Normal Pod vertx-create-span-sidecar-78898ff57-pws6c Binding Scheduled Successfully assigned kuttl-test-top-sawfly/vertx-create-span-sidecar-78898ff57-pws6c to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:05 +0000 UTC Normal Pod vertx-create-span-sidecar-78898ff57-pws6c AddedInterface Add eth0 [10.131.0.62/23] from ovn-kubernetes multus logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:05 +0000 UTC Normal Pod vertx-create-span-sidecar-78898ff57-pws6c.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:05 +0000 UTC Normal Pod vertx-create-span-sidecar-78898ff57-pws6c.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:05 +0000 UTC Normal Pod vertx-create-span-sidecar-78898ff57-pws6c.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:05 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-78898ff57 SuccessfulCreate Created pod: vertx-create-span-sidecar-78898ff57-pws6c replicaset-controller logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:05 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-78898ff57 to 1 deployment-controller logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:06 +0000 UTC Normal Pod vertx-create-span-sidecar-78898ff57-pws6c.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:06 +0000 UTC Normal Pod vertx-create-span-sidecar-78898ff57-pws6c.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:51:07 | sidecar-skip-webhook | 2025-01-27 07:51:06 +0000 UTC Normal Pod vertx-create-span-sidecar-78898ff57-pws6c.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:51:07 | sidecar-skip-webhook | Deleting namespace: kuttl-test-top-sawfly === CONT kuttl/harness/sidecar-deployment logger.go:42: 07:51:13 | sidecar-deployment | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:51:13 | sidecar-deployment | Creating namespace: kuttl-test-robust-warthog logger.go:42: 07:51:13 | sidecar-deployment/0-install | starting test step 0-install logger.go:42: 07:51:13 | sidecar-deployment/0-install | Jaeger:kuttl-test-robust-warthog/agent-as-sidecar created logger.go:42: 07:51:19 | sidecar-deployment/0-install | test step completed 0-install logger.go:42: 07:51:19 | sidecar-deployment/1-install | starting test step 1-install logger.go:42: 07:51:19 | sidecar-deployment/1-install | Deployment:kuttl-test-robust-warthog/vertx-create-span-sidecar created logger.go:42: 07:51:21 | sidecar-deployment/1-install | test step completed 1-install logger.go:42: 07:51:21 | sidecar-deployment/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:51:21 | sidecar-deployment/2-enable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-robust-warthog] logger.go:42: 07:51:21 | sidecar-deployment/2-enable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:51:23 | sidecar-deployment/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:51:23 | sidecar-deployment/3-find-service | starting test step 3-find-service logger.go:42: 07:51:23 | sidecar-deployment/3-find-service | Job:kuttl-test-robust-warthog/00-find-service created logger.go:42: 07:51:35 | sidecar-deployment/3-find-service | test step completed 3-find-service logger.go:42: 07:51:35 | sidecar-deployment/4-other-instance | starting test step 4-other-instance logger.go:42: 07:51:35 | sidecar-deployment/4-other-instance | Jaeger:kuttl-test-robust-warthog/agent-as-sidecar2 created logger.go:42: 07:51:41 | sidecar-deployment/4-other-instance | test step completed 4-other-instance logger.go:42: 07:51:41 | sidecar-deployment/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:51:42 | sidecar-deployment/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:51:42 | sidecar-deployment/6-find-service | starting test step 6-find-service logger.go:42: 07:51:42 | sidecar-deployment/6-find-service | Job:kuttl-test-robust-warthog/01-find-service created logger.go:42: 07:52:02 | sidecar-deployment/6-find-service | test step completed 6-find-service logger.go:42: 07:52:02 | sidecar-deployment/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:52:02 | sidecar-deployment/7-disable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=false --namespace kuttl-test-robust-warthog] logger.go:42: 07:52:02 | sidecar-deployment/7-disable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:52:05 | sidecar-deployment/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:52:05 | sidecar-deployment | sidecar-deployment events from ns kuttl-test-robust-warthog: logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:17 +0000 UTC Normal Pod agent-as-sidecar-5dcd9cb74b-stcjd Binding Scheduled Successfully assigned kuttl-test-robust-warthog/agent-as-sidecar-5dcd9cb74b-stcjd to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:17 +0000 UTC Normal Pod agent-as-sidecar-5dcd9cb74b-stcjd AddedInterface Add eth0 [10.129.2.48/23] from ovn-kubernetes multus logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:17 +0000 UTC Normal Pod agent-as-sidecar-5dcd9cb74b-stcjd.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:17 +0000 UTC Normal Pod agent-as-sidecar-5dcd9cb74b-stcjd.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:17 +0000 UTC Normal Pod agent-as-sidecar-5dcd9cb74b-stcjd.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:17 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-5dcd9cb74b SuccessfulCreate Created pod: agent-as-sidecar-5dcd9cb74b-stcjd replicaset-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:17 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-5dcd9cb74b to 1 deployment-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:19 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-k5tbq Binding Scheduled Successfully assigned kuttl-test-robust-warthog/vertx-create-span-sidecar-797645c8fc-k5tbq to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:19 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulCreate Created pod: vertx-create-span-sidecar-797645c8fc-k5tbq replicaset-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:19 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-797645c8fc to 1 deployment-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:20 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-k5tbq AddedInterface Add eth0 [10.131.0.63/23] from ovn-kubernetes multus logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:20 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-k5tbq.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:20 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-k5tbq.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:20 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-k5tbq.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:21 +0000 UTC Normal Pod vertx-create-span-sidecar-7685c8fb57-f2m72 Binding Scheduled Successfully assigned kuttl-test-robust-warthog/vertx-create-span-sidecar-7685c8fb57-f2m72 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:21 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-7685c8fb57 SuccessfulCreate Created pod: vertx-create-span-sidecar-7685c8fb57-f2m72 replicaset-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:21 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-7685c8fb57 to 1 deployment-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:22 +0000 UTC Normal Pod vertx-create-span-sidecar-7685c8fb57-f2m72 AddedInterface Add eth0 [10.128.2.69/23] from ovn-kubernetes multus logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:22 +0000 UTC Normal Pod vertx-create-span-sidecar-7685c8fb57-f2m72.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:22 +0000 UTC Normal Pod vertx-create-span-sidecar-7685c8fb57-f2m72.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:22 +0000 UTC Normal Pod vertx-create-span-sidecar-7685c8fb57-f2m72.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:22 +0000 UTC Normal Pod vertx-create-span-sidecar-7685c8fb57-f2m72.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:22 +0000 UTC Normal Pod vertx-create-span-sidecar-7685c8fb57-f2m72.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:22 +0000 UTC Normal Pod vertx-create-span-sidecar-7685c8fb57-f2m72.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:23 +0000 UTC Normal Pod 00-find-service-b6lq4 Binding Scheduled Successfully assigned kuttl-test-robust-warthog/00-find-service-b6lq4 to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:23 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-b6lq4 job-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:24 +0000 UTC Normal Pod 00-find-service-b6lq4 AddedInterface Add eth0 [10.129.2.49/23] from ovn-kubernetes multus logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:24 +0000 UTC Normal Pod 00-find-service-b6lq4.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:25 +0000 UTC Normal Pod 00-find-service-b6lq4.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 846ms (846ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:25 +0000 UTC Normal Pod 00-find-service-b6lq4.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:25 +0000 UTC Normal Pod 00-find-service-b6lq4.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:28 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-k5tbq.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.63:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:28 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-k5tbq.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.63:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:30 +0000 UTC Warning Pod vertx-create-span-sidecar-7685c8fb57-f2m72.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.69:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:30 +0000 UTC Warning Pod vertx-create-span-sidecar-7685c8fb57-f2m72.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.69:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:30 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-k5tbq.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:31 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-k5tbq.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.63:8080/": read tcp 10.131.0.2:45012->10.131.0.63:8080: read: connection reset by peer kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:31 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-k5tbq.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.63:8080/": dial tcp 10.131.0.63:8080: connect: connection refused kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:33 +0000 UTC Normal Pod vertx-create-span-sidecar-7685c8fb57-f2m72.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:33 +0000 UTC Warning Pod vertx-create-span-sidecar-7685c8fb57-f2m72.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.69:8080/": read tcp 10.128.2.2:38456->10.128.2.69:8080: read: connection reset by peer kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:33 +0000 UTC Warning Pod vertx-create-span-sidecar-7685c8fb57-f2m72.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.69:8080/": dial tcp 10.128.2.69:8080: connect: connection refused kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:35 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:39 +0000 UTC Normal Pod agent-as-sidecar2-5b65f49d4d-vm7lc Binding Scheduled Successfully assigned kuttl-test-robust-warthog/agent-as-sidecar2-5b65f49d4d-vm7lc to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:39 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-5b65f49d4d SuccessfulCreate Created pod: agent-as-sidecar2-5b65f49d4d-vm7lc replicaset-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:39 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-5b65f49d4d to 1 deployment-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:40 +0000 UTC Normal Pod agent-as-sidecar2-5b65f49d4d-vm7lc AddedInterface Add eth0 [10.129.2.50/23] from ovn-kubernetes multus logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:40 +0000 UTC Normal Pod agent-as-sidecar2-5b65f49d4d-vm7lc.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:40 +0000 UTC Normal Pod agent-as-sidecar2-5b65f49d4d-vm7lc.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:40 +0000 UTC Normal Pod agent-as-sidecar2-5b65f49d4d-vm7lc.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:41 +0000 UTC Normal Pod agent-as-sidecar-5dcd9cb74b-stcjd.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:42 +0000 UTC Normal Pod 01-find-service-jgl6f Binding Scheduled Successfully assigned kuttl-test-robust-warthog/01-find-service-jgl6f to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:42 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-jgl6f job-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:42 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-k5tbq.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.63:8080/": read tcp 10.131.0.2:58464->10.131.0.63:8080: read: connection reset by peer kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:43 +0000 UTC Normal Pod 01-find-service-jgl6f AddedInterface Add eth0 [10.129.2.51/23] from ovn-kubernetes multus logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:43 +0000 UTC Normal Pod 01-find-service-jgl6f.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:44 +0000 UTC Normal Pod 01-find-service-jgl6f.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 605ms (605ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:44 +0000 UTC Normal Pod 01-find-service-jgl6f.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:44 +0000 UTC Normal Pod 01-find-service-jgl6f.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:45 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulDelete Deleted pod: vertx-create-span-sidecar-797645c8fc-k5tbq replicaset-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:45 +0000 UTC Normal Pod vertx-create-span-sidecar-7dd858b6cf-8z97k Binding Scheduled Successfully assigned kuttl-test-robust-warthog/vertx-create-span-sidecar-7dd858b6cf-8z97k to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:45 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-7dd858b6cf SuccessfulCreate Created pod: vertx-create-span-sidecar-7dd858b6cf-8z97k replicaset-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:45 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-797645c8fc to 0 from 1 deployment-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:45 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-7dd858b6cf to 1 from 0 deployment-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:46 +0000 UTC Normal Pod vertx-create-span-sidecar-7dd858b6cf-8z97k AddedInterface Add eth0 [10.131.0.64/23] from ovn-kubernetes multus logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:46 +0000 UTC Normal Pod vertx-create-span-sidecar-7dd858b6cf-8z97k.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:46 +0000 UTC Normal Pod vertx-create-span-sidecar-7dd858b6cf-8z97k.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:46 +0000 UTC Normal Pod vertx-create-span-sidecar-7dd858b6cf-8z97k.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:46 +0000 UTC Normal Pod vertx-create-span-sidecar-7dd858b6cf-8z97k.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:46 +0000 UTC Normal Pod vertx-create-span-sidecar-7dd858b6cf-8z97k.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:46 +0000 UTC Normal Pod vertx-create-span-sidecar-7dd858b6cf-8z97k.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:54 +0000 UTC Warning Pod vertx-create-span-sidecar-7dd858b6cf-8z97k.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.64:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:54 +0000 UTC Warning Pod vertx-create-span-sidecar-7dd858b6cf-8z97k.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.64:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:56 +0000 UTC Normal Pod vertx-create-span-sidecar-7dd858b6cf-8z97k.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:57 +0000 UTC Warning Pod vertx-create-span-sidecar-7dd858b6cf-8z97k.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.64:8080/": read tcp 10.131.0.2:34496->10.131.0.64:8080: read: connection reset by peer kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:51:57 +0000 UTC Warning Pod vertx-create-span-sidecar-7dd858b6cf-8z97k.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.64:8080/": dial tcp 10.131.0.64:8080: connect: connection refused kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:52:02 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:52:02 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-7685c8fb57 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-7685c8fb57-f2m72 replicaset-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:52:02 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-7685c8fb57 to 0 from 1 deployment-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:52:02 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6c45c6c5b8 to 1 from 0 deployment-controller logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:52:03 +0000 UTC Normal Pod vertx-create-span-sidecar-6c45c6c5b8-fj429 Binding Scheduled Successfully assigned kuttl-test-robust-warthog/vertx-create-span-sidecar-6c45c6c5b8-fj429 to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:52:03 +0000 UTC Normal Pod vertx-create-span-sidecar-6c45c6c5b8-fj429 AddedInterface Add eth0 [10.129.2.52/23] from ovn-kubernetes multus logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:52:03 +0000 UTC Normal Pod vertx-create-span-sidecar-6c45c6c5b8-fj429.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:52:03 +0000 UTC Normal Pod vertx-create-span-sidecar-6c45c6c5b8-fj429.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:52:03 +0000 UTC Normal Pod vertx-create-span-sidecar-6c45c6c5b8-fj429.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:52:05 | sidecar-deployment | 2025-01-27 07:52:03 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6c45c6c5b8 SuccessfulCreate Created pod: vertx-create-span-sidecar-6c45c6c5b8-fj429 replicaset-controller logger.go:42: 07:52:05 | sidecar-deployment | Deleting namespace: kuttl-test-robust-warthog === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (146.37s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.96s) --- PASS: kuttl/harness/sidecar-namespace (65.80s) --- PASS: kuttl/harness/sidecar-skip-webhook (16.34s) --- PASS: kuttl/harness/sidecar-deployment (58.22s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name sidecar --report --output /logs/artifacts/sidecar.xml ./artifacts/kuttl-report.xml time="2025-01-27T07:52:11Z" level=debug msg="Setting a new name for the test suites" time="2025-01-27T07:52:11Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-01-27T07:52:11Z" level=debug msg="normalizing test case names" time="2025-01-27T07:52:11Z" level=debug msg="sidecar/artifacts -> sidecar_artifacts" time="2025-01-27T07:52:11Z" level=debug msg="sidecar/sidecar-namespace -> sidecar_sidecar_namespace" time="2025-01-27T07:52:11Z" level=debug msg="sidecar/sidecar-skip-webhook -> sidecar_sidecar_skip_webhook" time="2025-01-27T07:52:11Z" level=debug msg="sidecar/sidecar-deployment -> sidecar_sidecar_deployment" +------------------------------+--------+ | NAME | RESULT | +------------------------------+--------+ | sidecar_artifacts | passed | | sidecar_sidecar_namespace | passed | | sidecar_sidecar_skip_webhook | passed | | sidecar_sidecar_deployment | passed | +------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh streaming false true + '[' 3 -ne 3 ']' + test_suite_name=streaming + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/streaming.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-streaming make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ SKIP_ES_EXTERNAL=true \ ./tests/e2e/streaming/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-25-163410 True False 41m Cluster version is 4.18.0-0.nightly-2025-01-25-163410' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-25-163410 True False 41m Cluster version is 4.18.0-0.nightly-2025-01-25-163410' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/streaming/render.sh ++ export SUITE_DIR=./tests/e2e/streaming ++ SUITE_DIR=./tests/e2e/streaming ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/streaming ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + '[' false = true ']' + start_test streaming-simple + '[' 1 -ne 1 ']' + test_name=streaming-simple + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-simple' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-simple\e[0m' Rendering files for test streaming-simple + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + mkdir -p streaming-simple + cd streaming-simple + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/streaming-jaeger-assert.yaml.template -o ./04-assert.yaml + render_smoke_test simple-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=simple-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + export JAEGER_NAME=simple-streaming + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-tls + '[' 1 -ne 1 ']' + test_name=streaming-with-tls + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-tls' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-tls\e[0m' Rendering files for test streaming-with-tls + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-simple + '[' streaming-simple '!=' _build ']' + cd .. + mkdir -p streaming-with-tls + cd streaming-with-tls + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + render_smoke_test tls-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=tls-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + export JAEGER_NAME=tls-streaming + JAEGER_NAME=tls-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-autoprovisioning-autoscale + '[' 1 -ne 1 ']' + test_name=streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-autoprovisioning-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-autoprovisioning-autoscale\e[0m' Rendering files for test streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-tls + '[' streaming-with-tls '!=' _build ']' + cd .. + mkdir -p streaming-with-autoprovisioning-autoscale + cd streaming-with-autoprovisioning-autoscale + '[' true = true ']' + rm ./00-install.yaml ./00-assert.yaml + render_install_elasticsearch upstream 01 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=01 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./01-assert.yaml + jaeger_name=auto-provisioned + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="20Mi"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="500m"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.autoscale=true ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.minReplicas=1 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.maxReplicas=2 ./02-install.yaml + render_assert_kafka true auto-provisioned 03 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provisioned + test_step=03 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./03-assert.yaml ++ expr 03 + 1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./04-assert.yaml ++ expr 03 + 2 + CLUSTER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./05-assert.yaml + version_lt 1.30 1.23 ++ echo 1.30 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.30 '!=' 1.30 + rm ./08-assert.yaml + skip_test streaming-with-tls 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-with-tls + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-autoprovisioning-autoscale + '[' streaming-with-autoprovisioning-autoscale '!=' _build ']' + cd .. + rm -rf streaming-with-tls + warning 'streaming-with-tls: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-with-tls: This test is flaky in Prow CI\e[0m' WAR: streaming-with-tls: This test is flaky in Prow CI + skip_test streaming-simple 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-simple + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + rm -rf streaming-simple + warning 'streaming-simple: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-simple: This test is flaky in Prow CI\e[0m' WAR: streaming-simple: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running streaming E2E tests' Running streaming E2E tests + cd tests/e2e/streaming/_build + set +e + KUBECONFIG=/tmp/kubeconfig-4053090600 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 2 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/streaming-with-autoprovisioning-autoscale === PAUSE kuttl/harness/streaming-with-autoprovisioning-autoscale === CONT kuttl/harness/artifacts logger.go:42: 07:52:23 | artifacts | Creating namespace: kuttl-test-natural-hookworm logger.go:42: 07:52:23 | artifacts | artifacts events from ns kuttl-test-natural-hookworm: logger.go:42: 07:52:23 | artifacts | Deleting namespace: kuttl-test-natural-hookworm === CONT kuttl/harness/streaming-with-autoprovisioning-autoscale logger.go:42: 07:52:29 | streaming-with-autoprovisioning-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:52:29 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:52:29 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:52:29 | streaming-with-autoprovisioning-autoscale | Creating namespace: kuttl-test-settled-asp logger.go:42: 07:52:29 | streaming-with-autoprovisioning-autoscale/1-install | starting test step 1-install logger.go:42: 07:52:29 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc create sa deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:52:29 | streaming-with-autoprovisioning-autoscale/1-install | serviceaccount/deploy-elasticsearch created logger.go:42: 07:52:29 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc adm policy add-scc-to-user privileged -z deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:52:29 | streaming-with-autoprovisioning-autoscale/1-install | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:privileged added: "deploy-elasticsearch" logger.go:42: 07:52:29 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 6] logger.go:42: 07:52:35 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_0.yml -n $NAMESPACE] logger.go:42: 07:52:35 | streaming-with-autoprovisioning-autoscale/1-install | statefulset.apps/elasticsearch created logger.go:42: 07:52:35 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 3] logger.go:42: 07:52:38 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_1.yml -n $NAMESPACE] logger.go:42: 07:52:39 | streaming-with-autoprovisioning-autoscale/1-install | service/elasticsearch created logger.go:42: 07:52:57 | streaming-with-autoprovisioning-autoscale/1-install | test step completed 1-install logger.go:42: 07:52:57 | streaming-with-autoprovisioning-autoscale/2-install | starting test step 2-install logger.go:42: 07:52:57 | streaming-with-autoprovisioning-autoscale/2-install | Jaeger:kuttl-test-settled-asp/auto-provisioned created logger.go:42: 07:52:57 | streaming-with-autoprovisioning-autoscale/2-install | test step completed 2-install logger.go:42: 07:52:57 | streaming-with-autoprovisioning-autoscale/3- | starting test step 3- logger.go:42: 07:53:48 | streaming-with-autoprovisioning-autoscale/3- | test step completed 3- logger.go:42: 07:53:48 | streaming-with-autoprovisioning-autoscale/4- | starting test step 4- logger.go:42: 07:54:27 | streaming-with-autoprovisioning-autoscale/4- | test step completed 4- logger.go:42: 07:54:27 | streaming-with-autoprovisioning-autoscale/5- | starting test step 5- logger.go:42: 07:54:50 | streaming-with-autoprovisioning-autoscale/5- | test step completed 5- logger.go:42: 07:54:50 | streaming-with-autoprovisioning-autoscale/6- | starting test step 6- logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale/6- | test step completed 6- logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale/7- | starting test step 7- logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale/7- | test step completed 7- logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | streaming-with-autoprovisioning-autoscale events from ns kuttl-test-settled-asp: logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:52:35 +0000 UTC Normal Pod elasticsearch-0 Binding Scheduled Successfully assigned kuttl-test-settled-asp/elasticsearch-0 to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:52:35 +0000 UTC Normal StatefulSet.apps elasticsearch SuccessfulCreate create Pod elasticsearch-0 in StatefulSet elasticsearch successful statefulset-controller logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:52:36 +0000 UTC Normal Pod elasticsearch-0 AddedInterface Add eth0 [10.129.2.53/23] from ovn-kubernetes multus logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:52:36 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulling Pulling image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:52:47 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulled Successfully pulled image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" in 10.662s (10.662s including waiting). Image size: 758467647 bytes. kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:52:47 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:52:47 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:04 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-zookeeper NoPods No matching pods found controllermanager logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:04 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:04 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-settled-asp/data-auto-provisioned-zookeeper-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-b8cd7f56b-96fhj_2629f189-0baa-48cf-bb97-97063c2a37a2 logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:04 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:06 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ProvisioningSucceeded Successfully provisioned volume pvc-6b73c49b-d08c-41d3-ba58-fbc7053b7124 ebs.csi.aws.com_aws-ebs-csi-driver-controller-b8cd7f56b-96fhj_2629f189-0baa-48cf-bb97-97063c2a37a2 logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:07 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 Binding Scheduled Successfully assigned kuttl-test-settled-asp/auto-provisioned-zookeeper-0 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:10 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-6b73c49b-d08c-41d3-ba58-fbc7053b7124" attachdetach-controller logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:17 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 AddedInterface Add eth0 [10.128.2.70/23] from ovn-kubernetes multus logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:17 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:c9a60d03827466ee37dd4aff6803eda2c1d65cb2508cb57d13e1d47f982e5a20" kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:24 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:c9a60d03827466ee37dd4aff6803eda2c1d65cb2508cb57d13e1d47f982e5a20" in 7.185s (7.185s including waiting). Image size: 616797249 bytes. kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:25 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Created Created container zookeeper kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:25 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Started Started container zookeeper kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:49 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-kafka NoPods No matching pods found controllermanager logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:49 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:49 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-settled-asp/data-0-auto-provisioned-kafka-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-b8cd7f56b-96fhj_2629f189-0baa-48cf-bb97-97063c2a37a2 logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:49 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:51 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ProvisioningSucceeded Successfully provisioned volume pvc-6215b53d-87c1-44f6-a187-cd034fb570b0 ebs.csi.aws.com_aws-ebs-csi-driver-controller-b8cd7f56b-96fhj_2629f189-0baa-48cf-bb97-97063c2a37a2 logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:52 +0000 UTC Normal Pod auto-provisioned-kafka-0 Binding Scheduled Successfully assigned kuttl-test-settled-asp/auto-provisioned-kafka-0 to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:56 +0000 UTC Normal Pod auto-provisioned-kafka-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-6215b53d-87c1-44f6-a187-cd034fb570b0" attachdetach-controller logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:58 +0000 UTC Normal Pod auto-provisioned-kafka-0 AddedInterface Add eth0 [10.131.0.65/23] from ovn-kubernetes multus logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:53:58 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:c9a60d03827466ee37dd4aff6803eda2c1d65cb2508cb57d13e1d47f982e5a20" kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:04 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:c9a60d03827466ee37dd4aff6803eda2c1d65cb2508cb57d13e1d47f982e5a20" in 6.871s (6.871s including waiting). Image size: 616797249 bytes. kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:04 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Created Created container kafka kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:04 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Started Started container kafka kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:29 +0000 UTC Normal Pod auto-provisioned-entity-operator-6c47bf5f8b-npnjf Binding Scheduled Successfully assigned kuttl-test-settled-asp/auto-provisioned-entity-operator-6c47bf5f8b-npnjf to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:29 +0000 UTC Normal ReplicaSet.apps auto-provisioned-entity-operator-6c47bf5f8b SuccessfulCreate Created pod: auto-provisioned-entity-operator-6c47bf5f8b-npnjf replicaset-controller logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:29 +0000 UTC Normal Deployment.apps auto-provisioned-entity-operator ScalingReplicaSet Scaled up replica set auto-provisioned-entity-operator-6c47bf5f8b to 1 deployment-controller logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:30 +0000 UTC Normal Pod auto-provisioned-entity-operator-6c47bf5f8b-npnjf AddedInterface Add eth0 [10.129.2.54/23] from ovn-kubernetes multus logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:30 +0000 UTC Normal Pod auto-provisioned-entity-operator-6c47bf5f8b-npnjf.spec.containers{topic-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:07a9540a8b906c6d52e6a9684cfb838aadf9849cc24b1d80218ea3ad5545cb5a" already present on machine kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:30 +0000 UTC Normal Pod auto-provisioned-entity-operator-6c47bf5f8b-npnjf.spec.containers{topic-operator} Created Created container topic-operator kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:30 +0000 UTC Normal Pod auto-provisioned-entity-operator-6c47bf5f8b-npnjf.spec.containers{topic-operator} Started Started container topic-operator kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:30 +0000 UTC Normal Pod auto-provisioned-entity-operator-6c47bf5f8b-npnjf.spec.containers{user-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:07a9540a8b906c6d52e6a9684cfb838aadf9849cc24b1d80218ea3ad5545cb5a" already present on machine kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:30 +0000 UTC Normal Pod auto-provisioned-entity-operator-6c47bf5f8b-npnjf.spec.containers{user-operator} Created Created container user-operator kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:30 +0000 UTC Normal Pod auto-provisioned-entity-operator-6c47bf5f8b-npnjf.spec.containers{user-operator} Started Started container user-operator kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:53 +0000 UTC Normal Pod auto-provisioned-collector-846945477c-2dcfh Binding Scheduled Successfully assigned kuttl-test-settled-asp/auto-provisioned-collector-846945477c-2dcfh to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:53 +0000 UTC Normal Pod auto-provisioned-collector-846945477c-2dcfh AddedInterface Add eth0 [10.131.0.66/23] from ovn-kubernetes multus logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:53 +0000 UTC Normal Pod auto-provisioned-collector-846945477c-2dcfh.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:53 +0000 UTC Normal Pod auto-provisioned-collector-846945477c-2dcfh.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:53 +0000 UTC Normal Pod auto-provisioned-collector-846945477c-2dcfh.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:53 +0000 UTC Normal ReplicaSet.apps auto-provisioned-collector-846945477c SuccessfulCreate Created pod: auto-provisioned-collector-846945477c-2dcfh replicaset-controller logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:53 +0000 UTC Normal Deployment.apps auto-provisioned-collector ScalingReplicaSet Scaled up replica set auto-provisioned-collector-846945477c to 1 deployment-controller logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:53 +0000 UTC Normal Pod auto-provisioned-ingester-77bc5584db-4gskn Binding Scheduled Successfully assigned kuttl-test-settled-asp/auto-provisioned-ingester-77bc5584db-4gskn to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:53 +0000 UTC Normal Pod auto-provisioned-ingester-77bc5584db-4gskn AddedInterface Add eth0 [10.128.2.71/23] from ovn-kubernetes multus logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:53 +0000 UTC Normal Pod auto-provisioned-ingester-77bc5584db-4gskn.spec.containers{jaeger-ingester} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:e01a99dddffe65385d77dd6692558aa9a47df36f8b2a8d141b6ad561139981b2" kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:53 +0000 UTC Normal ReplicaSet.apps auto-provisioned-ingester-77bc5584db SuccessfulCreate Created pod: auto-provisioned-ingester-77bc5584db-4gskn replicaset-controller logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:53 +0000 UTC Normal Deployment.apps auto-provisioned-ingester ScalingReplicaSet Scaled up replica set auto-provisioned-ingester-77bc5584db to 1 deployment-controller logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:53 +0000 UTC Normal Pod auto-provisioned-query-648cd77746-95t6q Binding Scheduled Successfully assigned kuttl-test-settled-asp/auto-provisioned-query-648cd77746-95t6q to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:53 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-648cd77746 SuccessfulCreate Created pod: auto-provisioned-query-648cd77746-95t6q replicaset-controller logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:53 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled up replica set auto-provisioned-query-648cd77746 to 1 deployment-controller logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:54 +0000 UTC Normal Pod auto-provisioned-query-648cd77746-95t6q AddedInterface Add eth0 [10.128.2.72/23] from ovn-kubernetes multus logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:54 +0000 UTC Normal Pod auto-provisioned-query-648cd77746-95t6q.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:54 +0000 UTC Normal Pod auto-provisioned-query-648cd77746-95t6q.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:54 +0000 UTC Normal Pod auto-provisioned-query-648cd77746-95t6q.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:54 +0000 UTC Normal Pod auto-provisioned-query-648cd77746-95t6q.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:54 +0000 UTC Normal Pod auto-provisioned-query-648cd77746-95t6q.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:54 +0000 UTC Normal Pod auto-provisioned-query-648cd77746-95t6q.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:54 +0000 UTC Normal Pod auto-provisioned-query-648cd77746-95t6q.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:54 +0000 UTC Normal Pod auto-provisioned-query-648cd77746-95t6q.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:54 +0000 UTC Normal Pod auto-provisioned-query-648cd77746-95t6q.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:58 +0000 UTC Normal Pod auto-provisioned-ingester-77bc5584db-4gskn.spec.containers{jaeger-ingester} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:e01a99dddffe65385d77dd6692558aa9a47df36f8b2a8d141b6ad561139981b2" in 4.539s (4.539s including waiting). Image size: 139765359 bytes. kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:58 +0000 UTC Normal Pod auto-provisioned-ingester-77bc5584db-4gskn.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:58 +0000 UTC Normal Pod auto-provisioned-ingester-77bc5584db-4gskn.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | 2025-01-27 07:54:59 +0000 UTC Warning Pod auto-provisioned-ingester-77bc5584db-4gskn.spec.containers{jaeger-ingester} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 07:55:04 | streaming-with-autoprovisioning-autoscale | Deleting namespace: kuttl-test-settled-asp === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (178.61s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.87s) --- PASS: kuttl/harness/streaming-with-autoprovisioning-autoscale (172.70s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name streaming --report --output /logs/artifacts/streaming.xml ./artifacts/kuttl-report.xml time="2025-01-27T07:55:22Z" level=debug msg="Setting a new name for the test suites" time="2025-01-27T07:55:22Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-01-27T07:55:22Z" level=debug msg="normalizing test case names" time="2025-01-27T07:55:22Z" level=debug msg="streaming/artifacts -> streaming_artifacts" time="2025-01-27T07:55:22Z" level=debug msg="streaming/streaming-with-autoprovisioning-autoscale -> streaming_streaming_with_autoprovisioning_autoscale" +-----------------------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------------------+--------+ | streaming_artifacts | passed | | streaming_streaming_with_autoprovisioning_autoscale | passed | +-----------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh ui false true + '[' 3 -ne 3 ']' + test_suite_name=ui + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/ui.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-ui make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true ./tests/e2e/ui/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-25-163410 True False 44m Cluster version is 4.18.0-0.nightly-2025-01-25-163410' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-25-163410 True False 44m Cluster version is 4.18.0-0.nightly-2025-01-25-163410' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/ui/render.sh ++ export SUITE_DIR=./tests/e2e/ui ++ SUITE_DIR=./tests/e2e/ui ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/ui ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test allinone + '[' 1 -ne 1 ']' + test_name=allinone + echo =========================================================================== =========================================================================== + info 'Rendering files for test allinone' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test allinone\e[0m' Rendering files for test allinone + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build + '[' _build '!=' _build ']' + mkdir -p allinone + cd allinone + export GET_URL_COMMAND + export URL + export JAEGER_NAME=all-in-one-ui + JAEGER_NAME=all-in-one-ui + '[' true = true ']' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./01-curl.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./04-test-ui-config.yaml + start_test production + '[' 1 -ne 1 ']' + test_name=production + echo =========================================================================== =========================================================================== + info 'Rendering files for test production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test production\e[0m' Rendering files for test production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build/allinone + '[' allinone '!=' _build ']' + cd .. + mkdir -p production + cd production + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + [[ true = true ]] + [[ true = true ]] + render_install_jaeger production-ui production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + '[' true = true ']' + INSECURE=true + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-forbbiden-access.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-curl.yaml + INSECURE=true + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./05-check-disabled-security.yaml + ASSERT_PRESENT=false + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./06-check-NO-gaID.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./08-check-gaID.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running ui E2E tests' Running ui E2E tests + cd tests/e2e/ui/_build + set +e + KUBECONFIG=/tmp/kubeconfig-4053090600 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 3 tests === RUN kuttl/harness === RUN kuttl/harness/allinone === PAUSE kuttl/harness/allinone === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/production === PAUSE kuttl/harness/production === CONT kuttl/harness/allinone logger.go:42: 07:55:28 | allinone | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:55:28 | allinone | Creating namespace: kuttl-test-above-urchin logger.go:42: 07:55:28 | allinone/0-install | starting test step 0-install logger.go:42: 07:55:28 | allinone/0-install | Jaeger:kuttl-test-above-urchin/all-in-one-ui created logger.go:42: 07:55:32 | allinone/0-install | test step completed 0-install logger.go:42: 07:55:32 | allinone/1-curl | starting test step 1-curl logger.go:42: 07:55:32 | allinone/1-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 07:55:32 | allinone/1-curl | Checking the Ingress host value was populated logger.go:42: 07:55:32 | allinone/1-curl | Try number 0 logger.go:42: 07:55:32 | allinone/1-curl | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 07:55:32 | allinone/1-curl | template was: logger.go:42: 07:55:32 | allinone/1-curl | {.items[0].status.ingress[0].host} logger.go:42: 07:55:32 | allinone/1-curl | object given to jsonpath engine was: logger.go:42: 07:55:32 | allinone/1-curl | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 07:55:32 | allinone/1-curl | logger.go:42: 07:55:32 | allinone/1-curl | logger.go:42: 07:55:42 | allinone/1-curl | Try number 1 logger.go:42: 07:55:43 | allinone/1-curl | Hostname is all-in-one-ui-kuttl-test-above-urchin.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:55:43 | allinone/1-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE all-in-one-ui] logger.go:42: 07:55:43 | allinone/1-curl | Checking an expected HTTP response logger.go:42: 07:55:43 | allinone/1-curl | Running in OpenShift logger.go:42: 07:55:43 | allinone/1-curl | User not provided. Getting the token... logger.go:42: 07:55:44 | allinone/1-curl | Warning: resource jaegers/all-in-one-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:55:50 | allinone/1-curl | Try number 1/30 the https://all-in-one-ui-kuttl-test-above-urchin.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:55:50 | allinone/1-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:55:50 | allinone/1-curl | Try number 2/30 the https://all-in-one-ui-kuttl-test-above-urchin.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:55:50 | allinone/1-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 07:56:00 | allinone/1-curl | Try number 3/30 the https://all-in-one-ui-kuttl-test-above-urchin.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:56:00 | allinone/1-curl | curl response asserted properly logger.go:42: 07:56:00 | allinone/1-curl | test step completed 1-curl logger.go:42: 07:56:00 | allinone/2-delete | starting test step 2-delete logger.go:42: 07:56:00 | allinone/2-delete | Jaeger:kuttl-test-above-urchin/all-in-one-ui created logger.go:42: 07:56:00 | allinone/2-delete | test step completed 2-delete logger.go:42: 07:56:00 | allinone/3-install | starting test step 3-install logger.go:42: 07:56:00 | allinone/3-install | Jaeger:kuttl-test-above-urchin/all-in-one-ui updated logger.go:42: 07:56:00 | allinone/3-install | test step completed 3-install logger.go:42: 07:56:00 | allinone/4-test-ui-config | starting test step 4-test-ui-config logger.go:42: 07:56:00 | allinone/4-test-ui-config | running command: [./ensure-ingress-host.sh] logger.go:42: 07:56:00 | allinone/4-test-ui-config | Checking the Ingress host value was populated logger.go:42: 07:56:00 | allinone/4-test-ui-config | Try number 0 logger.go:42: 07:56:00 | allinone/4-test-ui-config | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 07:56:00 | allinone/4-test-ui-config | template was: logger.go:42: 07:56:00 | allinone/4-test-ui-config | {.items[0].status.ingress[0].host} logger.go:42: 07:56:00 | allinone/4-test-ui-config | object given to jsonpath engine was: logger.go:42: 07:56:00 | allinone/4-test-ui-config | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 07:56:00 | allinone/4-test-ui-config | logger.go:42: 07:56:00 | allinone/4-test-ui-config | logger.go:42: 07:56:10 | allinone/4-test-ui-config | Try number 1 logger.go:42: 07:56:10 | allinone/4-test-ui-config | Hostname is all-in-one-ui-kuttl-test-above-urchin.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:56:10 | allinone/4-test-ui-config | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:56:11 | allinone/4-test-ui-config | time="2025-01-27T07:56:11Z" level=info msg="Querying https://all-in-one-ui-kuttl-test-above-urchin.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:56:11 | allinone/4-test-ui-config | time="2025-01-27T07:56:11Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:56:11 | allinone/4-test-ui-config | time="2025-01-27T07:56:11Z" level=info msg="Polling to https://all-in-one-ui-kuttl-test-above-urchin.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:56:11 | allinone/4-test-ui-config | time="2025-01-27T07:56:11Z" level=info msg="Doing request number 0" logger.go:42: 07:56:11 | allinone/4-test-ui-config | time="2025-01-27T07:56:11Z" level=info msg="Content found and asserted!" logger.go:42: 07:56:11 | allinone/4-test-ui-config | time="2025-01-27T07:56:11Z" level=info msg="Success!" logger.go:42: 07:56:11 | allinone/4-test-ui-config | test step completed 4-test-ui-config logger.go:42: 07:56:11 | allinone | allinone events from ns kuttl-test-above-urchin: logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:32 +0000 UTC Normal Pod all-in-one-ui-58794444d6-khrtk Binding Scheduled Successfully assigned kuttl-test-above-urchin/all-in-one-ui-58794444d6-khrtk to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:32 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-58794444d6 SuccessfulCreate Created pod: all-in-one-ui-58794444d6-khrtk replicaset-controller logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:32 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-58794444d6 to 1 deployment-controller logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:33 +0000 UTC Normal Pod all-in-one-ui-58794444d6-khrtk AddedInterface Add eth0 [10.128.2.73/23] from ovn-kubernetes multus logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:33 +0000 UTC Normal Pod all-in-one-ui-58794444d6-khrtk.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:33 +0000 UTC Normal Pod all-in-one-ui-58794444d6-khrtk.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:33 +0000 UTC Normal Pod all-in-one-ui-58794444d6-khrtk.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:33 +0000 UTC Normal Pod all-in-one-ui-58794444d6-khrtk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:33 +0000 UTC Normal Pod all-in-one-ui-58794444d6-khrtk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:33 +0000 UTC Normal Pod all-in-one-ui-58794444d6-khrtk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:45 +0000 UTC Normal Pod all-in-one-ui-58794444d6-khrtk.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:45 +0000 UTC Normal Pod all-in-one-ui-58794444d6-khrtk.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:45 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-58794444d6 SuccessfulDelete Deleted pod: all-in-one-ui-58794444d6-khrtk replicaset-controller logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:45 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-58794444d6 to 0 from 1 deployment-controller logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:46 +0000 UTC Normal Pod all-in-one-ui-db876ccf4-tdhb8 Binding Scheduled Successfully assigned kuttl-test-above-urchin/all-in-one-ui-db876ccf4-tdhb8 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:46 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-db876ccf4 SuccessfulCreate Created pod: all-in-one-ui-db876ccf4-tdhb8 replicaset-controller logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:46 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-db876ccf4 to 1 deployment-controller logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:47 +0000 UTC Normal Pod all-in-one-ui-db876ccf4-tdhb8 AddedInterface Add eth0 [10.128.2.74/23] from ovn-kubernetes multus logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:47 +0000 UTC Normal Pod all-in-one-ui-db876ccf4-tdhb8.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:47 +0000 UTC Normal Pod all-in-one-ui-db876ccf4-tdhb8.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:47 +0000 UTC Normal Pod all-in-one-ui-db876ccf4-tdhb8.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:47 +0000 UTC Normal Pod all-in-one-ui-db876ccf4-tdhb8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:47 +0000 UTC Normal Pod all-in-one-ui-db876ccf4-tdhb8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:56:11 | allinone | 2025-01-27 07:55:47 +0000 UTC Normal Pod all-in-one-ui-db876ccf4-tdhb8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:56:11 | allinone | 2025-01-27 07:56:00 +0000 UTC Normal Pod all-in-one-ui-db876ccf4-tdhb8.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:56:11 | allinone | 2025-01-27 07:56:00 +0000 UTC Normal Pod all-in-one-ui-db876ccf4-tdhb8.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:56:11 | allinone | 2025-01-27 07:56:07 +0000 UTC Normal Pod all-in-one-ui-748b8b8958-mc4lt Binding Scheduled Successfully assigned kuttl-test-above-urchin/all-in-one-ui-748b8b8958-mc4lt to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:56:11 | allinone | 2025-01-27 07:56:07 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-748b8b8958 SuccessfulCreate Created pod: all-in-one-ui-748b8b8958-mc4lt replicaset-controller logger.go:42: 07:56:11 | allinone | 2025-01-27 07:56:07 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-748b8b8958 to 1 deployment-controller logger.go:42: 07:56:11 | allinone | 2025-01-27 07:56:08 +0000 UTC Normal Pod all-in-one-ui-748b8b8958-mc4lt AddedInterface Add eth0 [10.129.2.55/23] from ovn-kubernetes multus logger.go:42: 07:56:11 | allinone | 2025-01-27 07:56:08 +0000 UTC Normal Pod all-in-one-ui-748b8b8958-mc4lt.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:56:11 | allinone | 2025-01-27 07:56:08 +0000 UTC Normal Pod all-in-one-ui-748b8b8958-mc4lt.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:56:11 | allinone | 2025-01-27 07:56:08 +0000 UTC Normal Pod all-in-one-ui-748b8b8958-mc4lt.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:56:11 | allinone | Deleting namespace: kuttl-test-above-urchin === CONT kuttl/harness/production logger.go:42: 07:56:17 | production | Ignoring add-tracking-id.yaml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:56:17 | production | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:56:17 | production | Creating namespace: kuttl-test-needed-lioness logger.go:42: 07:56:17 | production/1-install | starting test step 1-install logger.go:42: 07:56:17 | production/1-install | Jaeger:kuttl-test-needed-lioness/production-ui created logger.go:42: 07:56:54 | production/1-install | test step completed 1-install logger.go:42: 07:56:54 | production/2-check-forbbiden-access | starting test step 2-check-forbbiden-access logger.go:42: 07:56:54 | production/2-check-forbbiden-access | running command: [./ensure-ingress-host.sh] logger.go:42: 07:56:54 | production/2-check-forbbiden-access | Checking the Ingress host value was populated logger.go:42: 07:56:54 | production/2-check-forbbiden-access | Try number 0 logger.go:42: 07:56:54 | production/2-check-forbbiden-access | Hostname is production-ui-kuttl-test-needed-lioness.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:56:54 | production/2-check-forbbiden-access | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE production-ui] logger.go:42: 07:56:54 | production/2-check-forbbiden-access | Checking an expected HTTP response logger.go:42: 07:56:54 | production/2-check-forbbiden-access | Running in OpenShift logger.go:42: 07:56:54 | production/2-check-forbbiden-access | Not using any secret logger.go:42: 07:56:54 | production/2-check-forbbiden-access | Try number 1/30 the https://production-ui-kuttl-test-needed-lioness.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:56:54 | production/2-check-forbbiden-access | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:56:54 | production/2-check-forbbiden-access | Try number 2/30 the https://production-ui-kuttl-test-needed-lioness.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:56:54 | production/2-check-forbbiden-access | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 07:57:04 | production/2-check-forbbiden-access | Try number 3/30 the https://production-ui-kuttl-test-needed-lioness.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:57:04 | production/2-check-forbbiden-access | curl response asserted properly logger.go:42: 07:57:04 | production/2-check-forbbiden-access | test step completed 2-check-forbbiden-access logger.go:42: 07:57:04 | production/3-curl | starting test step 3-curl logger.go:42: 07:57:04 | production/3-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 07:57:04 | production/3-curl | Checking the Ingress host value was populated logger.go:42: 07:57:04 | production/3-curl | Try number 0 logger.go:42: 07:57:04 | production/3-curl | Hostname is production-ui-kuttl-test-needed-lioness.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:57:04 | production/3-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 07:57:05 | production/3-curl | Checking an expected HTTP response logger.go:42: 07:57:05 | production/3-curl | Running in OpenShift logger.go:42: 07:57:05 | production/3-curl | User not provided. Getting the token... logger.go:42: 07:57:06 | production/3-curl | Warning: resource jaegers/production-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:57:12 | production/3-curl | Try number 1/30 the https://production-ui-kuttl-test-needed-lioness.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:57:12 | production/3-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:57:12 | production/3-curl | Try number 2/30 the https://production-ui-kuttl-test-needed-lioness.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:57:12 | production/3-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 07:57:22 | production/3-curl | Try number 3/30 the https://production-ui-kuttl-test-needed-lioness.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:57:22 | production/3-curl | curl response asserted properly logger.go:42: 07:57:22 | production/3-curl | test step completed 3-curl logger.go:42: 07:57:22 | production/4-install | starting test step 4-install logger.go:42: 07:57:22 | production/4-install | Jaeger:kuttl-test-needed-lioness/production-ui updated logger.go:42: 07:57:22 | production/4-install | test step completed 4-install logger.go:42: 07:57:22 | production/5-check-disabled-security | starting test step 5-check-disabled-security logger.go:42: 07:57:22 | production/5-check-disabled-security | running command: [./ensure-ingress-host.sh] logger.go:42: 07:57:22 | production/5-check-disabled-security | Checking the Ingress host value was populated logger.go:42: 07:57:22 | production/5-check-disabled-security | Try number 0 logger.go:42: 07:57:22 | production/5-check-disabled-security | Hostname is production-ui-kuttl-test-needed-lioness.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:57:22 | production/5-check-disabled-security | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 07:57:22 | production/5-check-disabled-security | Checking an expected HTTP response logger.go:42: 07:57:22 | production/5-check-disabled-security | Running in OpenShift logger.go:42: 07:57:22 | production/5-check-disabled-security | Not using any secret logger.go:42: 07:57:22 | production/5-check-disabled-security | Try number 1/30 the https://production-ui-kuttl-test-needed-lioness.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:57:22 | production/5-check-disabled-security | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:57:22 | production/5-check-disabled-security | Try number 2/30 the https://production-ui-kuttl-test-needed-lioness.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:57:22 | production/5-check-disabled-security | HTTP response is 403. 200 expected. Waiting 10 s logger.go:42: 07:57:32 | production/5-check-disabled-security | Try number 3/30 the https://production-ui-kuttl-test-needed-lioness.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:57:32 | production/5-check-disabled-security | curl response asserted properly logger.go:42: 07:57:32 | production/5-check-disabled-security | test step completed 5-check-disabled-security logger.go:42: 07:57:32 | production/6-check-NO-gaID | starting test step 6-check-NO-gaID logger.go:42: 07:57:32 | production/6-check-NO-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 07:57:32 | production/6-check-NO-gaID | Checking the Ingress host value was populated logger.go:42: 07:57:32 | production/6-check-NO-gaID | Try number 0 logger.go:42: 07:57:32 | production/6-check-NO-gaID | Hostname is production-ui-kuttl-test-needed-lioness.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:57:32 | production/6-check-NO-gaID | running command: [sh -c ASSERT_PRESENT=false EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:57:33 | production/6-check-NO-gaID | time="2025-01-27T07:57:33Z" level=info msg="Querying https://production-ui-kuttl-test-needed-lioness.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:57:33 | production/6-check-NO-gaID | time="2025-01-27T07:57:33Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:57:33 | production/6-check-NO-gaID | time="2025-01-27T07:57:33Z" level=info msg="Polling to https://production-ui-kuttl-test-needed-lioness.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:57:33 | production/6-check-NO-gaID | time="2025-01-27T07:57:33Z" level=info msg="Doing request number 0" logger.go:42: 07:57:33 | production/6-check-NO-gaID | time="2025-01-27T07:57:33Z" level=info msg="Content not found and asserted it was not found!" logger.go:42: 07:57:33 | production/6-check-NO-gaID | time="2025-01-27T07:57:33Z" level=info msg="Success!" logger.go:42: 07:57:33 | production/6-check-NO-gaID | test step completed 6-check-NO-gaID logger.go:42: 07:57:33 | production/7-add-tracking-id | starting test step 7-add-tracking-id logger.go:42: 07:57:33 | production/7-add-tracking-id | running command: [sh -c kubectl apply -f add-tracking-id.yaml -n $NAMESPACE] logger.go:42: 07:57:33 | production/7-add-tracking-id | jaeger.jaegertracing.io/production-ui configured logger.go:42: 07:57:33 | production/7-add-tracking-id | test step completed 7-add-tracking-id logger.go:42: 07:57:33 | production/8-check-gaID | starting test step 8-check-gaID logger.go:42: 07:57:33 | production/8-check-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 07:57:33 | production/8-check-gaID | Checking the Ingress host value was populated logger.go:42: 07:57:33 | production/8-check-gaID | Try number 0 logger.go:42: 07:57:33 | production/8-check-gaID | Hostname is production-ui-kuttl-test-needed-lioness.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:57:33 | production/8-check-gaID | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:57:34 | production/8-check-gaID | time="2025-01-27T07:57:34Z" level=info msg="Querying https://production-ui-kuttl-test-needed-lioness.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:57:34 | production/8-check-gaID | time="2025-01-27T07:57:34Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:57:34 | production/8-check-gaID | time="2025-01-27T07:57:34Z" level=info msg="Polling to https://production-ui-kuttl-test-needed-lioness.apps.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:57:34 | production/8-check-gaID | time="2025-01-27T07:57:34Z" level=info msg="Doing request number 0" logger.go:42: 07:57:34 | production/8-check-gaID | time="2025-01-27T07:57:34Z" level=warning msg="Found: false . Assert: true" logger.go:42: 07:57:34 | production/8-check-gaID | time="2025-01-27T07:57:34Z" level=warning msg="The condition of the test function was not accomplished" logger.go:42: 07:57:42 | production/8-check-gaID | time="2025-01-27T07:57:42Z" level=info msg="Doing request number 1" logger.go:42: 07:58:02 | production/8-check-gaID | time="2025-01-27T07:58:02Z" level=info msg="Content found and asserted!" logger.go:42: 07:58:02 | production/8-check-gaID | time="2025-01-27T07:58:02Z" level=info msg="Success!" logger.go:42: 07:58:02 | production/8-check-gaID | test step completed 8-check-gaID logger.go:42: 07:58:02 | production | production events from ns kuttl-test-needed-lioness: logger.go:42: 07:58:02 | production | 2025-01-27 07:56:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestneededlionessproductionui-1-5795m8jk Binding Scheduled Successfully assigned kuttl-test-needed-lioness/elasticsearch-cdm-kuttltestneededlionessproductionui-1-5795m8jk to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:02 | production | 2025-01-27 07:56:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestneededlionessproductionui-1-5795m8jk AddedInterface Add eth0 [10.129.2.56/23] from ovn-kubernetes multus logger.go:42: 07:58:02 | production | 2025-01-27 07:56:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestneededlionessproductionui-1-5795m8jk.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestneededlionessproductionui-1-5795m8jk.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestneededlionessproductionui-1-5795m8jk.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestneededlionessproductionui-1-5795m8jk.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestneededlionessproductionui-1-5795m8jk.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestneededlionessproductionui-1-5795m8jk.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:24 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestneededlionessproductionui-1-5796bb474c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestneededlionessproductionui-1-5795m8jk replicaset-controller logger.go:42: 07:58:02 | production | 2025-01-27 07:56:24 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestneededlionessproductionui-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestneededlionessproductionui-1-5796bb474c to 1 deployment-controller logger.go:42: 07:58:02 | production | 2025-01-27 07:56:34 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestneededlionessproductionui-1-5795m8jk.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:39 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestneededlionessproductionui-1-5795m8jk.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:50 +0000 UTC Normal Pod production-ui-collector-6f464ff687-7gm5p Binding Scheduled Successfully assigned kuttl-test-needed-lioness/production-ui-collector-6f464ff687-7gm5p to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:02 | production | 2025-01-27 07:56:50 +0000 UTC Normal ReplicaSet.apps production-ui-collector-6f464ff687 SuccessfulCreate Created pod: production-ui-collector-6f464ff687-7gm5p replicaset-controller logger.go:42: 07:58:02 | production | 2025-01-27 07:56:50 +0000 UTC Normal Deployment.apps production-ui-collector ScalingReplicaSet Scaled up replica set production-ui-collector-6f464ff687 to 1 deployment-controller logger.go:42: 07:58:02 | production | 2025-01-27 07:56:50 +0000 UTC Normal Pod production-ui-query-6d7cd4774-vv9df Binding Scheduled Successfully assigned kuttl-test-needed-lioness/production-ui-query-6d7cd4774-vv9df to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:02 | production | 2025-01-27 07:56:50 +0000 UTC Normal ReplicaSet.apps production-ui-query-6d7cd4774 SuccessfulCreate Created pod: production-ui-query-6d7cd4774-vv9df replicaset-controller logger.go:42: 07:58:02 | production | 2025-01-27 07:56:50 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-6d7cd4774 to 1 deployment-controller logger.go:42: 07:58:02 | production | 2025-01-27 07:56:51 +0000 UTC Normal Pod production-ui-collector-6f464ff687-7gm5p AddedInterface Add eth0 [10.128.2.75/23] from ovn-kubernetes multus logger.go:42: 07:58:02 | production | 2025-01-27 07:56:51 +0000 UTC Normal Pod production-ui-collector-6f464ff687-7gm5p.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:51 +0000 UTC Normal Pod production-ui-collector-6f464ff687-7gm5p.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:51 +0000 UTC Normal Pod production-ui-collector-6f464ff687-7gm5p.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:51 +0000 UTC Normal Pod production-ui-query-6d7cd4774-vv9df AddedInterface Add eth0 [10.131.0.67/23] from ovn-kubernetes multus logger.go:42: 07:58:02 | production | 2025-01-27 07:56:51 +0000 UTC Normal Pod production-ui-query-6d7cd4774-vv9df.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:51 +0000 UTC Normal Pod production-ui-query-6d7cd4774-vv9df.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:51 +0000 UTC Normal Pod production-ui-query-6d7cd4774-vv9df.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:51 +0000 UTC Normal Pod production-ui-query-6d7cd4774-vv9df.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:51 +0000 UTC Normal Pod production-ui-query-6d7cd4774-vv9df.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:51 +0000 UTC Normal Pod production-ui-query-6d7cd4774-vv9df.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:51 +0000 UTC Normal Pod production-ui-query-6d7cd4774-vv9df.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:51 +0000 UTC Normal Pod production-ui-query-6d7cd4774-vv9df.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:56:51 +0000 UTC Normal Pod production-ui-query-6d7cd4774-vv9df.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:06 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:58:02 | production | 2025-01-27 07:57:06 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:58:02 | production | 2025-01-27 07:57:06 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:58:02 | production | 2025-01-27 07:57:08 +0000 UTC Normal Pod production-ui-query-599c7596b7-dvgj7 Binding Scheduled Successfully assigned kuttl-test-needed-lioness/production-ui-query-599c7596b7-dvgj7 to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:02 | production | 2025-01-27 07:57:08 +0000 UTC Normal ReplicaSet.apps production-ui-query-599c7596b7 SuccessfulCreate Created pod: production-ui-query-599c7596b7-dvgj7 replicaset-controller logger.go:42: 07:58:02 | production | 2025-01-27 07:57:08 +0000 UTC Normal Pod production-ui-query-6d7cd4774-vv9df.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:08 +0000 UTC Normal Pod production-ui-query-6d7cd4774-vv9df.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:08 +0000 UTC Normal Pod production-ui-query-6d7cd4774-vv9df.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:08 +0000 UTC Normal ReplicaSet.apps production-ui-query-6d7cd4774 SuccessfulDelete Deleted pod: production-ui-query-6d7cd4774-vv9df replicaset-controller logger.go:42: 07:58:02 | production | 2025-01-27 07:57:08 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-6d7cd4774 to 0 from 1 deployment-controller logger.go:42: 07:58:02 | production | 2025-01-27 07:57:08 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-599c7596b7 to 1 deployment-controller logger.go:42: 07:58:02 | production | 2025-01-27 07:57:09 +0000 UTC Normal Pod production-ui-query-599c7596b7-dvgj7 AddedInterface Add eth0 [10.131.0.68/23] from ovn-kubernetes multus logger.go:42: 07:58:02 | production | 2025-01-27 07:57:09 +0000 UTC Normal Pod production-ui-query-599c7596b7-dvgj7.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:09 +0000 UTC Normal Pod production-ui-query-599c7596b7-dvgj7.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:09 +0000 UTC Normal Pod production-ui-query-599c7596b7-dvgj7.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:09 +0000 UTC Normal Pod production-ui-query-599c7596b7-dvgj7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:09 +0000 UTC Normal Pod production-ui-query-599c7596b7-dvgj7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:09 +0000 UTC Normal Pod production-ui-query-599c7596b7-dvgj7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:09 +0000 UTC Normal Pod production-ui-query-599c7596b7-dvgj7.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:09 +0000 UTC Normal Pod production-ui-query-599c7596b7-dvgj7.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:09 +0000 UTC Normal Pod production-ui-query-599c7596b7-dvgj7.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:24 +0000 UTC Normal Pod production-ui-query-599c7596b7-dvgj7.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:24 +0000 UTC Normal Pod production-ui-query-599c7596b7-dvgj7.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:24 +0000 UTC Normal Pod production-ui-query-599c7596b7-dvgj7.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:24 +0000 UTC Normal ReplicaSet.apps production-ui-query-599c7596b7 SuccessfulDelete Deleted pod: production-ui-query-599c7596b7-dvgj7 replicaset-controller logger.go:42: 07:58:02 | production | 2025-01-27 07:57:24 +0000 UTC Normal Pod production-ui-query-c8df98c67-9s7p8 Binding Scheduled Successfully assigned kuttl-test-needed-lioness/production-ui-query-c8df98c67-9s7p8 to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:02 | production | 2025-01-27 07:57:24 +0000 UTC Normal ReplicaSet.apps production-ui-query-c8df98c67 SuccessfulCreate Created pod: production-ui-query-c8df98c67-9s7p8 replicaset-controller logger.go:42: 07:58:02 | production | 2025-01-27 07:57:24 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-599c7596b7 to 0 from 1 deployment-controller logger.go:42: 07:58:02 | production | 2025-01-27 07:57:24 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-c8df98c67 to 1 deployment-controller logger.go:42: 07:58:02 | production | 2025-01-27 07:57:25 +0000 UTC Normal Pod production-ui-query-c8df98c67-9s7p8 AddedInterface Add eth0 [10.131.0.69/23] from ovn-kubernetes multus logger.go:42: 07:58:02 | production | 2025-01-27 07:57:25 +0000 UTC Normal Pod production-ui-query-c8df98c67-9s7p8.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:25 +0000 UTC Normal Pod production-ui-query-c8df98c67-9s7p8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:25 +0000 UTC Normal Pod production-ui-query-c8df98c67-9s7p8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:25 +0000 UTC Normal Pod production-ui-query-c8df98c67-9s7p8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:25 +0000 UTC Normal Pod production-ui-query-c8df98c67-9s7p8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:25 +0000 UTC Normal Pod production-ui-query-c8df98c67-9s7p8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:35 +0000 UTC Normal Pod production-ui-query-88899dfd6-d5qfz Binding Scheduled Successfully assigned kuttl-test-needed-lioness/production-ui-query-88899dfd6-d5qfz to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:02 | production | 2025-01-27 07:57:35 +0000 UTC Normal ReplicaSet.apps production-ui-query-88899dfd6 SuccessfulCreate Created pod: production-ui-query-88899dfd6-d5qfz replicaset-controller logger.go:42: 07:58:02 | production | 2025-01-27 07:57:35 +0000 UTC Normal Pod production-ui-query-c8df98c67-9s7p8.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:35 +0000 UTC Normal Pod production-ui-query-c8df98c67-9s7p8.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:35 +0000 UTC Warning Pod production-ui-query-c8df98c67-9s7p8.spec.containers{jaeger-agent} Unhealthy Readiness probe failed: Get "http://10.131.0.69:14271/": dial tcp 10.131.0.69:14271: connect: connection refused kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:35 +0000 UTC Warning Pod production-ui-query-c8df98c67-9s7p8.spec.containers{jaeger-query} Unhealthy Readiness probe failed: Get "http://10.131.0.69:16687/": dial tcp 10.131.0.69:16687: connect: connection refused kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:35 +0000 UTC Normal ReplicaSet.apps production-ui-query-c8df98c67 SuccessfulDelete Deleted pod: production-ui-query-c8df98c67-9s7p8 replicaset-controller logger.go:42: 07:58:02 | production | 2025-01-27 07:57:35 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-c8df98c67 to 0 from 1 deployment-controller logger.go:42: 07:58:02 | production | 2025-01-27 07:57:35 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-88899dfd6 to 1 deployment-controller logger.go:42: 07:58:02 | production | 2025-01-27 07:57:36 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod production-ui-collector-6f464ff687-7gm5p horizontal-pod-autoscaler logger.go:42: 07:58:02 | production | 2025-01-27 07:57:36 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod production-ui-collector-6f464ff687-7gm5p horizontal-pod-autoscaler logger.go:42: 07:58:02 | production | 2025-01-27 07:57:36 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod production-ui-collector-6f464ff687-7gm5p horizontal-pod-autoscaler logger.go:42: 07:58:02 | production | 2025-01-27 07:57:36 +0000 UTC Normal Pod production-ui-query-88899dfd6-d5qfz AddedInterface Add eth0 [10.131.0.70/23] from ovn-kubernetes multus logger.go:42: 07:58:02 | production | 2025-01-27 07:57:36 +0000 UTC Normal Pod production-ui-query-88899dfd6-d5qfz.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:36 +0000 UTC Normal Pod production-ui-query-88899dfd6-d5qfz.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:36 +0000 UTC Normal Pod production-ui-query-88899dfd6-d5qfz.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:36 +0000 UTC Normal Pod production-ui-query-88899dfd6-d5qfz.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:36 +0000 UTC Normal Pod production-ui-query-88899dfd6-d5qfz.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:58:02 | production | 2025-01-27 07:57:36 +0000 UTC Normal Pod production-ui-query-88899dfd6-d5qfz.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:58:02 | production | Deleting namespace: kuttl-test-needed-lioness === CONT kuttl/harness/artifacts logger.go:42: 07:58:08 | artifacts | Creating namespace: kuttl-test-gorgeous-killdeer logger.go:42: 07:58:08 | artifacts | artifacts events from ns kuttl-test-gorgeous-killdeer: logger.go:42: 07:58:08 | artifacts | Deleting namespace: kuttl-test-gorgeous-killdeer === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (165.78s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/allinone (48.38s) --- PASS: kuttl/harness/production (111.27s) --- PASS: kuttl/harness/artifacts (6.07s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name ui --report --output /logs/artifacts/ui.xml ./artifacts/kuttl-report.xml time="2025-01-27T07:58:14Z" level=debug msg="Setting a new name for the test suites" time="2025-01-27T07:58:14Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-01-27T07:58:14Z" level=debug msg="normalizing test case names" time="2025-01-27T07:58:14Z" level=debug msg="ui/allinone -> ui_allinone" time="2025-01-27T07:58:14Z" level=debug msg="ui/production -> ui_production" time="2025-01-27T07:58:14Z" level=debug msg="ui/artifacts -> ui_artifacts" +---------------+--------+ | NAME | RESULT | +---------------+--------+ | ui_allinone | passed | | ui_production | passed | | ui_artifacts | passed | +---------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh miscellaneous false true + '[' 3 -ne 3 ']' + test_suite_name=miscellaneous + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/miscellaneous.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-miscellaneous make[2]: Entering directory '/tmp/jaeger-tests' SKIP_ES_EXTERNAL=true ./tests/e2e/miscellaneous/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-25-163410 True False 47m Cluster version is 4.18.0-0.nightly-2025-01-25-163410' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-25-163410 True False 47m Cluster version is 4.18.0-0.nightly-2025-01-25-163410' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/miscellaneous/render.sh ++ export SUITE_DIR=./tests/e2e/miscellaneous ++ SUITE_DIR=./tests/e2e/miscellaneous ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/miscellaneous ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test collector-autoscale + '[' 1 -ne 1 ']' + test_name=collector-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-autoscale\e[0m' Rendering files for test collector-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p collector-autoscale + cd collector-autoscale + jaeger_name=simple-prod + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + ELASTICSEARCH_NODECOUNT=1 + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.resources.requests.memory="200m"' 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.autoscale=true 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.minReplicas=1 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.maxReplicas=2 01-install.yaml + version_lt 1.30 1.23 ++ echo 1.30 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.30 '!=' 1.30 + rm ./03-assert.yaml + generate_otlp_e2e_tests http + test_protocol=http + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-http\e[0m' Rendering files for test collector-otlp-allinone-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-autoscale + '[' collector-autoscale '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-http + cd collector-otlp-allinone-http + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger http true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-http\e[0m' Rendering files for test collector-otlp-production-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-http + '[' collector-otlp-allinone-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-http + cd collector-otlp-production-http + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger http true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + generate_otlp_e2e_tests grpc + test_protocol=grpc + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-grpc\e[0m' Rendering files for test collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-http + '[' collector-otlp-production-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-grpc + cd collector-otlp-allinone-grpc + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-grpc\e[0m' Rendering files for test collector-otlp-production-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-grpc + '[' collector-otlp-allinone-grpc '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-grpc + cd collector-otlp-production-grpc + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + '[' true = true ']' + skip_test istio 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=istio + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-grpc + '[' collector-otlp-production-grpc '!=' _build ']' + cd .. + rm -rf istio + warning 'istio: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: istio: Test not supported in OpenShift\e[0m' WAR: istio: Test not supported in OpenShift + '[' true = true ']' + skip_test outside-cluster 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=outside-cluster + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf outside-cluster + warning 'outside-cluster: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: outside-cluster: Test not supported in OpenShift\e[0m' WAR: outside-cluster: Test not supported in OpenShift + start_test set-custom-img + '[' 1 -ne 1 ']' + test_name=set-custom-img + echo =========================================================================== =========================================================================== + info 'Rendering files for test set-custom-img' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test set-custom-img\e[0m' Rendering files for test set-custom-img + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p set-custom-img + cd set-custom-img + jaeger_name=my-jaeger + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.image="test"' ./02-install.yaml + '[' true = true ']' + skip_test non-cluster-wide 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=non-cluster-wide + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/set-custom-img + '[' set-custom-img '!=' _build ']' + cd .. + rm -rf non-cluster-wide + warning 'non-cluster-wide: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: non-cluster-wide: Test not supported in OpenShift\e[0m' WAR: non-cluster-wide: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running miscellaneous E2E tests' Running miscellaneous E2E tests + cd tests/e2e/miscellaneous/_build + set +e + KUBECONFIG=/tmp/kubeconfig-4053090600 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-z5skswmd-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 8 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/cassandra-spark === PAUSE kuttl/harness/cassandra-spark === RUN kuttl/harness/collector-autoscale === PAUSE kuttl/harness/collector-autoscale === RUN kuttl/harness/collector-otlp-allinone-grpc === PAUSE kuttl/harness/collector-otlp-allinone-grpc === RUN kuttl/harness/collector-otlp-allinone-http === PAUSE kuttl/harness/collector-otlp-allinone-http === RUN kuttl/harness/collector-otlp-production-grpc === PAUSE kuttl/harness/collector-otlp-production-grpc === RUN kuttl/harness/collector-otlp-production-http === PAUSE kuttl/harness/collector-otlp-production-http === RUN kuttl/harness/set-custom-img === PAUSE kuttl/harness/set-custom-img === CONT kuttl/harness/artifacts logger.go:42: 07:58:25 | artifacts | Creating namespace: kuttl-test-viable-griffon logger.go:42: 07:58:25 | artifacts | artifacts events from ns kuttl-test-viable-griffon: logger.go:42: 07:58:25 | artifacts | Deleting namespace: kuttl-test-viable-griffon === CONT kuttl/harness/collector-otlp-allinone-http logger.go:42: 07:58:31 | collector-otlp-allinone-http | Creating namespace: kuttl-test-fit-narwhal logger.go:42: 07:58:31 | collector-otlp-allinone-http/0-install | starting test step 0-install logger.go:42: 07:58:31 | collector-otlp-allinone-http/0-install | Jaeger:kuttl-test-fit-narwhal/my-jaeger created logger.go:42: 07:58:37 | collector-otlp-allinone-http/0-install | test step completed 0-install logger.go:42: 07:58:37 | collector-otlp-allinone-http/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:58:37 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:58:38 | collector-otlp-allinone-http/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:58:44 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:58:45 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:58:45 | collector-otlp-allinone-http/1-smoke-test | job.batch/report-span created logger.go:42: 07:58:45 | collector-otlp-allinone-http/1-smoke-test | job.batch/check-span created logger.go:42: 07:58:57 | collector-otlp-allinone-http/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:58:57 | collector-otlp-allinone-http | collector-otlp-allinone-http events from ns kuttl-test-fit-narwhal: logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:35 +0000 UTC Normal Pod my-jaeger-7455b874d6-s8xbc Binding Scheduled Successfully assigned kuttl-test-fit-narwhal/my-jaeger-7455b874d6-s8xbc to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:35 +0000 UTC Normal ReplicaSet.apps my-jaeger-7455b874d6 SuccessfulCreate Created pod: my-jaeger-7455b874d6-s8xbc replicaset-controller logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:35 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-7455b874d6 to 1 deployment-controller logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:36 +0000 UTC Normal Pod my-jaeger-7455b874d6-s8xbc AddedInterface Add eth0 [10.128.2.76/23] from ovn-kubernetes multus logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:36 +0000 UTC Normal Pod my-jaeger-7455b874d6-s8xbc.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:36 +0000 UTC Normal Pod my-jaeger-7455b874d6-s8xbc.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:36 +0000 UTC Normal Pod my-jaeger-7455b874d6-s8xbc.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:36 +0000 UTC Normal Pod my-jaeger-7455b874d6-s8xbc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:36 +0000 UTC Normal Pod my-jaeger-7455b874d6-s8xbc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:36 +0000 UTC Normal Pod my-jaeger-7455b874d6-s8xbc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:41 +0000 UTC Normal Pod my-jaeger-7455b874d6-s8xbc.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:41 +0000 UTC Normal Pod my-jaeger-7455b874d6-s8xbc.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:41 +0000 UTC Normal ReplicaSet.apps my-jaeger-7455b874d6 SuccessfulDelete Deleted pod: my-jaeger-7455b874d6-s8xbc replicaset-controller logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:41 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-7455b874d6 to 0 from 1 deployment-controller logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:42 +0000 UTC Normal Pod my-jaeger-c8758c954-gztrr Binding Scheduled Successfully assigned kuttl-test-fit-narwhal/my-jaeger-c8758c954-gztrr to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:42 +0000 UTC Normal Pod my-jaeger-c8758c954-gztrr AddedInterface Add eth0 [10.128.2.77/23] from ovn-kubernetes multus logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:42 +0000 UTC Normal Pod my-jaeger-c8758c954-gztrr.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:42 +0000 UTC Normal Pod my-jaeger-c8758c954-gztrr.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:42 +0000 UTC Normal Pod my-jaeger-c8758c954-gztrr.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:42 +0000 UTC Normal Pod my-jaeger-c8758c954-gztrr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:42 +0000 UTC Normal Pod my-jaeger-c8758c954-gztrr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:42 +0000 UTC Normal Pod my-jaeger-c8758c954-gztrr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:42 +0000 UTC Normal ReplicaSet.apps my-jaeger-c8758c954 SuccessfulCreate Created pod: my-jaeger-c8758c954-gztrr replicaset-controller logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:42 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-c8758c954 to 1 deployment-controller logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:45 +0000 UTC Normal Pod check-span-6cgqb Binding Scheduled Successfully assigned kuttl-test-fit-narwhal/check-span-6cgqb to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:45 +0000 UTC Normal Pod check-span-6cgqb AddedInterface Add eth0 [10.131.0.71/23] from ovn-kubernetes multus logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:45 +0000 UTC Normal Pod check-span-6cgqb.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:45 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-6cgqb job-controller logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:45 +0000 UTC Normal Pod report-span-tzpnr Binding Scheduled Successfully assigned kuttl-test-fit-narwhal/report-span-tzpnr to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:45 +0000 UTC Normal Pod report-span-tzpnr AddedInterface Add eth0 [10.129.2.57/23] from ovn-kubernetes multus logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:45 +0000 UTC Normal Pod report-span-tzpnr.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:45 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-tzpnr job-controller logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:46 +0000 UTC Normal Pod check-span-6cgqb.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 641ms (641ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:46 +0000 UTC Normal Pod check-span-6cgqb.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:46 +0000 UTC Normal Pod check-span-6cgqb.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:46 +0000 UTC Normal Pod report-span-tzpnr.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.057s (1.057s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:46 +0000 UTC Normal Pod report-span-tzpnr.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:46 +0000 UTC Normal Pod report-span-tzpnr.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:58:57 | collector-otlp-allinone-http | 2025-01-27 07:58:57 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:58:57 | collector-otlp-allinone-http | Deleting namespace: kuttl-test-fit-narwhal === CONT kuttl/harness/set-custom-img logger.go:42: 07:59:09 | set-custom-img | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:59:09 | set-custom-img | Ignoring check-collector-img.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:59:09 | set-custom-img | Creating namespace: kuttl-test-robust-snail logger.go:42: 07:59:09 | set-custom-img/1-install | starting test step 1-install logger.go:42: 07:59:09 | set-custom-img/1-install | Jaeger:kuttl-test-robust-snail/my-jaeger created logger.go:42: 07:59:44 | set-custom-img/1-install | test step completed 1-install logger.go:42: 07:59:44 | set-custom-img/2-install | starting test step 2-install logger.go:42: 07:59:44 | set-custom-img/2-install | Jaeger:kuttl-test-robust-snail/my-jaeger updated logger.go:42: 07:59:44 | set-custom-img/2-install | test step completed 2-install logger.go:42: 07:59:44 | set-custom-img/3-check-image | starting test step 3-check-image logger.go:42: 07:59:44 | set-custom-img/3-check-image | running command: [sh -c ./check-collector-img.sh] logger.go:42: 07:59:44 | set-custom-img/3-check-image | Collector image missmatch. Expected: test. Has: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d logger.go:42: 07:59:49 | set-custom-img/3-check-image | Collector image asserted properly! logger.go:42: 07:59:49 | set-custom-img/3-check-image | test step completed 3-check-image logger.go:42: 07:59:49 | set-custom-img | set-custom-img events from ns kuttl-test-robust-snail: logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:15 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestrobustsnailmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestrobustsnailmyjaeger-1-66dbf686dc to 1 deployment-controller logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:16 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestrobustsnailmyjaeger-1-66dbf686dc SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestrobustsnailmyjaeger-1-66dbf686dnnf4g replicaset-controller logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrobustsnailmyjaeger-1-66dbf686dnnf4g Binding Scheduled Successfully assigned kuttl-test-robust-snail/elasticsearch-cdm-kuttltestrobustsnailmyjaeger-1-66dbf686dnnf4g to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrobustsnailmyjaeger-1-66dbf686dnnf4g AddedInterface Add eth0 [10.129.2.58/23] from ovn-kubernetes multus logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrobustsnailmyjaeger-1-66dbf686dnnf4g.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrobustsnailmyjaeger-1-66dbf686dnnf4g.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrobustsnailmyjaeger-1-66dbf686dnnf4g.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrobustsnailmyjaeger-1-66dbf686dnnf4g.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrobustsnailmyjaeger-1-66dbf686dnnf4g.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrobustsnailmyjaeger-1-66dbf686dnnf4g.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:26 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestrobustsnailmyjaeger-1-66dbf686dnnf4g.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:31 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestrobustsnailmyjaeger-1-66dbf686dnnf4g.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:42 +0000 UTC Normal Pod my-jaeger-collector-6995c4fd67-9x52k Binding Scheduled Successfully assigned kuttl-test-robust-snail/my-jaeger-collector-6995c4fd67-9x52k to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:42 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-6995c4fd67 SuccessfulCreate Created pod: my-jaeger-collector-6995c4fd67-9x52k replicaset-controller logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:42 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-6995c4fd67 to 1 deployment-controller logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:42 +0000 UTC Normal Pod my-jaeger-query-6b44688694-k4h5n Binding Scheduled Successfully assigned kuttl-test-robust-snail/my-jaeger-query-6b44688694-k4h5n to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:42 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-6b44688694 SuccessfulCreate Created pod: my-jaeger-query-6b44688694-k4h5n replicaset-controller logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:42 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-6b44688694 to 1 deployment-controller logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:43 +0000 UTC Normal Pod my-jaeger-collector-6995c4fd67-9x52k AddedInterface Add eth0 [10.131.0.72/23] from ovn-kubernetes multus logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:43 +0000 UTC Normal Pod my-jaeger-collector-6995c4fd67-9x52k.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:43 +0000 UTC Normal Pod my-jaeger-collector-6995c4fd67-9x52k.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:43 +0000 UTC Normal Pod my-jaeger-collector-6995c4fd67-9x52k.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:43 +0000 UTC Normal Pod my-jaeger-query-6b44688694-k4h5n AddedInterface Add eth0 [10.128.2.78/23] from ovn-kubernetes multus logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:43 +0000 UTC Normal Pod my-jaeger-query-6b44688694-k4h5n.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:43 +0000 UTC Normal Pod my-jaeger-query-6b44688694-k4h5n.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:43 +0000 UTC Normal Pod my-jaeger-query-6b44688694-k4h5n.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:43 +0000 UTC Normal Pod my-jaeger-query-6b44688694-k4h5n.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:43 +0000 UTC Normal Pod my-jaeger-query-6b44688694-k4h5n.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:43 +0000 UTC Normal Pod my-jaeger-query-6b44688694-k4h5n.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:43 +0000 UTC Normal Pod my-jaeger-query-6b44688694-k4h5n.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:43 +0000 UTC Normal Pod my-jaeger-query-6b44688694-k4h5n.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:43 +0000 UTC Normal Pod my-jaeger-query-6b44688694-k4h5n.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:49 +0000 UTC Normal Pod my-jaeger-collector-6995c4fd67-9x52k.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:49 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-6995c4fd67 SuccessfulDelete Deleted pod: my-jaeger-collector-6995c4fd67-9x52k replicaset-controller logger.go:42: 07:59:49 | set-custom-img | 2025-01-27 07:59:49 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-6995c4fd67 to 0 from 1 deployment-controller logger.go:42: 07:59:49 | set-custom-img | Deleting namespace: kuttl-test-robust-snail === CONT kuttl/harness/collector-otlp-production-http logger.go:42: 07:59:55 | collector-otlp-production-http | Creating namespace: kuttl-test-brief-poodle logger.go:42: 07:59:56 | collector-otlp-production-http/1-install | starting test step 1-install logger.go:42: 07:59:56 | collector-otlp-production-http/1-install | Jaeger:kuttl-test-brief-poodle/my-jaeger created logger.go:42: 08:00:34 | collector-otlp-production-http/1-install | test step completed 1-install logger.go:42: 08:00:34 | collector-otlp-production-http/2-smoke-test | starting test step 2-smoke-test logger.go:42: 08:00:34 | collector-otlp-production-http/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 08:00:35 | collector-otlp-production-http/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:00:41 | collector-otlp-production-http/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 08:00:42 | collector-otlp-production-http/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:00:42 | collector-otlp-production-http/2-smoke-test | job.batch/report-span created logger.go:42: 08:00:42 | collector-otlp-production-http/2-smoke-test | job.batch/check-span created logger.go:42: 08:00:56 | collector-otlp-production-http/2-smoke-test | test step completed 2-smoke-test logger.go:42: 08:00:56 | collector-otlp-production-http | collector-otlp-production-http events from ns kuttl-test-brief-poodle: logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:03 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestbriefpoodlemyjaeger-1-59d45fb45c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestbriefpoodlemyjaeger-1-59d45fb45xprfm replicaset-controller logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbriefpoodlemyjaeger-1-59d45fb45xprfm Binding Scheduled Successfully assigned kuttl-test-brief-poodle/elasticsearch-cdm-kuttltestbriefpoodlemyjaeger-1-59d45fb45xprfm to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:03 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestbriefpoodlemyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestbriefpoodlemyjaeger-1-59d45fb45c to 1 deployment-controller logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbriefpoodlemyjaeger-1-59d45fb45xprfm AddedInterface Add eth0 [10.129.2.59/23] from ovn-kubernetes multus logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbriefpoodlemyjaeger-1-59d45fb45xprfm.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbriefpoodlemyjaeger-1-59d45fb45xprfm.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbriefpoodlemyjaeger-1-59d45fb45xprfm.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbriefpoodlemyjaeger-1-59d45fb45xprfm.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbriefpoodlemyjaeger-1-59d45fb45xprfm.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbriefpoodlemyjaeger-1-59d45fb45xprfm.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:14 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestbriefpoodlemyjaeger-1-59d45fb45xprfm.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:19 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestbriefpoodlemyjaeger-1-59d45fb45xprfm.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:30 +0000 UTC Normal Pod my-jaeger-collector-5cb9fdb8f8-vppt7 Binding Scheduled Successfully assigned kuttl-test-brief-poodle/my-jaeger-collector-5cb9fdb8f8-vppt7 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:30 +0000 UTC Normal Pod my-jaeger-collector-5cb9fdb8f8-vppt7 AddedInterface Add eth0 [10.128.2.79/23] from ovn-kubernetes multus logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:30 +0000 UTC Normal Pod my-jaeger-collector-5cb9fdb8f8-vppt7.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:30 +0000 UTC Normal Pod my-jaeger-collector-5cb9fdb8f8-vppt7.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:30 +0000 UTC Normal Pod my-jaeger-collector-5cb9fdb8f8-vppt7.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:30 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5cb9fdb8f8 SuccessfulCreate Created pod: my-jaeger-collector-5cb9fdb8f8-vppt7 replicaset-controller logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:30 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-5cb9fdb8f8 to 1 deployment-controller logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:30 +0000 UTC Normal Pod my-jaeger-query-5f4445767d-t9ftq Binding Scheduled Successfully assigned kuttl-test-brief-poodle/my-jaeger-query-5f4445767d-t9ftq to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:30 +0000 UTC Warning Pod my-jaeger-query-5f4445767d-t9ftq FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-oauth-proxy-tls" : secret "my-jaeger-ui-oauth-proxy-tls" not found kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:30 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5f4445767d SuccessfulCreate Created pod: my-jaeger-query-5f4445767d-t9ftq replicaset-controller logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:30 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-5f4445767d to 1 deployment-controller logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:31 +0000 UTC Normal Pod my-jaeger-query-5f4445767d-t9ftq AddedInterface Add eth0 [10.131.0.74/23] from ovn-kubernetes multus logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:31 +0000 UTC Normal Pod my-jaeger-query-5f4445767d-t9ftq.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:31 +0000 UTC Normal Pod my-jaeger-query-5f4445767d-t9ftq.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:31 +0000 UTC Normal Pod my-jaeger-query-5f4445767d-t9ftq.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:31 +0000 UTC Normal Pod my-jaeger-query-5f4445767d-t9ftq.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:31 +0000 UTC Normal Pod my-jaeger-query-5f4445767d-t9ftq.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:31 +0000 UTC Normal Pod my-jaeger-query-5f4445767d-t9ftq.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:31 +0000 UTC Normal Pod my-jaeger-query-5f4445767d-t9ftq.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:31 +0000 UTC Normal Pod my-jaeger-query-5f4445767d-t9ftq.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:31 +0000 UTC Normal Pod my-jaeger-query-5f4445767d-t9ftq.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:37 +0000 UTC Normal Pod my-jaeger-query-5f4445767d-t9ftq.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:37 +0000 UTC Normal Pod my-jaeger-query-5f4445767d-t9ftq.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:37 +0000 UTC Normal Pod my-jaeger-query-5f4445767d-t9ftq.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:37 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5f4445767d SuccessfulDelete Deleted pod: my-jaeger-query-5f4445767d-t9ftq replicaset-controller logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:37 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-5f4445767d to 0 from 1 deployment-controller logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:38 +0000 UTC Normal Pod my-jaeger-query-5bc85f565b-r6v4j Binding Scheduled Successfully assigned kuttl-test-brief-poodle/my-jaeger-query-5bc85f565b-r6v4j to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:38 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5bc85f565b SuccessfulCreate Created pod: my-jaeger-query-5bc85f565b-r6v4j replicaset-controller logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:38 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-5bc85f565b to 1 deployment-controller logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:39 +0000 UTC Normal Pod my-jaeger-query-5bc85f565b-r6v4j AddedInterface Add eth0 [10.131.0.75/23] from ovn-kubernetes multus logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:39 +0000 UTC Normal Pod my-jaeger-query-5bc85f565b-r6v4j.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:39 +0000 UTC Normal Pod my-jaeger-query-5bc85f565b-r6v4j.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:39 +0000 UTC Normal Pod my-jaeger-query-5bc85f565b-r6v4j.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:39 +0000 UTC Normal Pod my-jaeger-query-5bc85f565b-r6v4j.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:39 +0000 UTC Normal Pod my-jaeger-query-5bc85f565b-r6v4j.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:39 +0000 UTC Normal Pod my-jaeger-query-5bc85f565b-r6v4j.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:39 +0000 UTC Normal Pod my-jaeger-query-5bc85f565b-r6v4j.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:39 +0000 UTC Normal Pod my-jaeger-query-5bc85f565b-r6v4j.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:39 +0000 UTC Normal Pod my-jaeger-query-5bc85f565b-r6v4j.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:42 +0000 UTC Normal Pod check-span-r7ctg Binding Scheduled Successfully assigned kuttl-test-brief-poodle/check-span-r7ctg to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:42 +0000 UTC Normal Pod check-span-r7ctg AddedInterface Add eth0 [10.128.2.81/23] from ovn-kubernetes multus logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:42 +0000 UTC Normal Pod check-span-r7ctg.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:42 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-r7ctg job-controller logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:42 +0000 UTC Normal Pod report-span-9x9mw Binding Scheduled Successfully assigned kuttl-test-brief-poodle/report-span-9x9mw to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:42 +0000 UTC Normal Pod report-span-9x9mw AddedInterface Add eth0 [10.128.2.80/23] from ovn-kubernetes multus logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:42 +0000 UTC Normal Pod report-span-9x9mw.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:42 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-9x9mw job-controller logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:43 +0000 UTC Normal Pod report-span-9x9mw.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 788ms (788ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:43 +0000 UTC Normal Pod report-span-9x9mw.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:43 +0000 UTC Normal Pod report-span-9x9mw.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:44 +0000 UTC Normal Pod check-span-r7ctg.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.651s (1.651s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:44 +0000 UTC Normal Pod check-span-r7ctg.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:44 +0000 UTC Normal Pod check-span-r7ctg.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:00:56 | collector-otlp-production-http | 2025-01-27 08:00:55 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:00:56 | collector-otlp-production-http | Deleting namespace: kuttl-test-brief-poodle === CONT kuttl/harness/collector-otlp-production-grpc logger.go:42: 08:01:08 | collector-otlp-production-grpc | Creating namespace: kuttl-test-pleased-pangolin logger.go:42: 08:01:08 | collector-otlp-production-grpc/1-install | starting test step 1-install logger.go:42: 08:01:08 | collector-otlp-production-grpc/1-install | Jaeger:kuttl-test-pleased-pangolin/my-jaeger created logger.go:42: 08:01:43 | collector-otlp-production-grpc/1-install | test step completed 1-install logger.go:42: 08:01:43 | collector-otlp-production-grpc/2-smoke-test | starting test step 2-smoke-test logger.go:42: 08:01:43 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 08:01:44 | collector-otlp-production-grpc/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:01:51 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 08:01:51 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:01:51 | collector-otlp-production-grpc/2-smoke-test | job.batch/report-span created logger.go:42: 08:01:51 | collector-otlp-production-grpc/2-smoke-test | job.batch/check-span created logger.go:42: 08:02:13 | collector-otlp-production-grpc/2-smoke-test | test step completed 2-smoke-test logger.go:42: 08:02:13 | collector-otlp-production-grpc | collector-otlp-production-grpc events from ns kuttl-test-pleased-pangolin: logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:14 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestpleasedpangolinmyjaeger-1-7fcdf999b7 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestpleasedpangolinmyjaeger-1-7fcdf9n9q2 replicaset-controller logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpleasedpangolinmyjaeger-1-7fcdf9n9q2 Binding Scheduled Successfully assigned kuttl-test-pleased-pangolin/elasticsearch-cdm-kuttltestpleasedpangolinmyjaeger-1-7fcdf9n9q2 to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpleasedpangolinmyjaeger-1-7fcdf9n9q2 AddedInterface Add eth0 [10.129.2.60/23] from ovn-kubernetes multus logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpleasedpangolinmyjaeger-1-7fcdf9n9q2.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpleasedpangolinmyjaeger-1-7fcdf9n9q2.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpleasedpangolinmyjaeger-1-7fcdf9n9q2.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpleasedpangolinmyjaeger-1-7fcdf9n9q2.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpleasedpangolinmyjaeger-1-7fcdf9n9q2.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpleasedpangolinmyjaeger-1-7fcdf9n9q2.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:14 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestpleasedpangolinmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestpleasedpangolinmyjaeger-1-7fcdf999b7 to 1 deployment-controller logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:24 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpleasedpangolinmyjaeger-1-7fcdf9n9q2.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:29 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpleasedpangolinmyjaeger-1-7fcdf9n9q2.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:41 +0000 UTC Normal Pod my-jaeger-collector-76677bc979-98tx8 Binding Scheduled Successfully assigned kuttl-test-pleased-pangolin/my-jaeger-collector-76677bc979-98tx8 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:41 +0000 UTC Warning Pod my-jaeger-collector-76677bc979-98tx8 FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:41 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-76677bc979 SuccessfulCreate Created pod: my-jaeger-collector-76677bc979-98tx8 replicaset-controller logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:41 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-76677bc979 to 1 deployment-controller logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:41 +0000 UTC Normal Pod my-jaeger-query-6ffdf4b686-sqwls Binding Scheduled Successfully assigned kuttl-test-pleased-pangolin/my-jaeger-query-6ffdf4b686-sqwls to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:41 +0000 UTC Normal Pod my-jaeger-query-6ffdf4b686-sqwls AddedInterface Add eth0 [10.131.0.76/23] from ovn-kubernetes multus logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:41 +0000 UTC Normal Pod my-jaeger-query-6ffdf4b686-sqwls.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:41 +0000 UTC Normal Pod my-jaeger-query-6ffdf4b686-sqwls.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:41 +0000 UTC Normal Pod my-jaeger-query-6ffdf4b686-sqwls.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:41 +0000 UTC Normal Pod my-jaeger-query-6ffdf4b686-sqwls.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:41 +0000 UTC Normal Pod my-jaeger-query-6ffdf4b686-sqwls.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:41 +0000 UTC Normal Pod my-jaeger-query-6ffdf4b686-sqwls.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:41 +0000 UTC Normal Pod my-jaeger-query-6ffdf4b686-sqwls.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:41 +0000 UTC Normal Pod my-jaeger-query-6ffdf4b686-sqwls.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:41 +0000 UTC Normal Pod my-jaeger-query-6ffdf4b686-sqwls.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:41 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-6ffdf4b686 SuccessfulCreate Created pod: my-jaeger-query-6ffdf4b686-sqwls replicaset-controller logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:41 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-6ffdf4b686 to 1 deployment-controller logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:42 +0000 UTC Normal Pod my-jaeger-collector-76677bc979-98tx8 AddedInterface Add eth0 [10.128.2.82/23] from ovn-kubernetes multus logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:42 +0000 UTC Normal Pod my-jaeger-collector-76677bc979-98tx8.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:42 +0000 UTC Normal Pod my-jaeger-collector-76677bc979-98tx8.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:42 +0000 UTC Normal Pod my-jaeger-collector-76677bc979-98tx8.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:48 +0000 UTC Normal Pod my-jaeger-query-6ffdf4b686-sqwls.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:48 +0000 UTC Normal Pod my-jaeger-query-6ffdf4b686-sqwls.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:48 +0000 UTC Normal Pod my-jaeger-query-6ffdf4b686-sqwls.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:48 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-6ffdf4b686 SuccessfulDelete Deleted pod: my-jaeger-query-6ffdf4b686-sqwls replicaset-controller logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:48 +0000 UTC Normal Pod my-jaeger-query-789c5f67b7-h2xl8 Binding Scheduled Successfully assigned kuttl-test-pleased-pangolin/my-jaeger-query-789c5f67b7-h2xl8 to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:48 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-789c5f67b7 SuccessfulCreate Created pod: my-jaeger-query-789c5f67b7-h2xl8 replicaset-controller logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:48 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-6ffdf4b686 to 0 from 1 deployment-controller logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:48 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-789c5f67b7 to 1 deployment-controller logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:49 +0000 UTC Normal Pod my-jaeger-query-789c5f67b7-h2xl8 AddedInterface Add eth0 [10.131.0.77/23] from ovn-kubernetes multus logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:49 +0000 UTC Normal Pod my-jaeger-query-789c5f67b7-h2xl8.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:49 +0000 UTC Normal Pod my-jaeger-query-789c5f67b7-h2xl8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:49 +0000 UTC Normal Pod my-jaeger-query-789c5f67b7-h2xl8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:49 +0000 UTC Normal Pod my-jaeger-query-789c5f67b7-h2xl8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:49 +0000 UTC Normal Pod my-jaeger-query-789c5f67b7-h2xl8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:49 +0000 UTC Normal Pod my-jaeger-query-789c5f67b7-h2xl8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:49 +0000 UTC Normal Pod my-jaeger-query-789c5f67b7-h2xl8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:49 +0000 UTC Normal Pod my-jaeger-query-789c5f67b7-h2xl8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:49 +0000 UTC Normal Pod my-jaeger-query-789c5f67b7-h2xl8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:51 +0000 UTC Normal Pod check-span-skqwz Binding Scheduled Successfully assigned kuttl-test-pleased-pangolin/check-span-skqwz to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:51 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-skqwz job-controller logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:51 +0000 UTC Normal Pod report-span-t2262 Binding Scheduled Successfully assigned kuttl-test-pleased-pangolin/report-span-t2262 to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:51 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-t2262 job-controller logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:52 +0000 UTC Normal Pod check-span-skqwz AddedInterface Add eth0 [10.128.2.84/23] from ovn-kubernetes multus logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:52 +0000 UTC Normal Pod check-span-skqwz.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:52 +0000 UTC Normal Pod report-span-t2262 AddedInterface Add eth0 [10.128.2.83/23] from ovn-kubernetes multus logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:52 +0000 UTC Normal Pod report-span-t2262.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:53 +0000 UTC Normal Pod report-span-t2262.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.639s (1.639s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:53 +0000 UTC Normal Pod report-span-t2262.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:53 +0000 UTC Normal Pod report-span-t2262.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:54 +0000 UTC Normal Pod check-span-skqwz.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.052s (2.052s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:54 +0000 UTC Normal Pod check-span-skqwz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:54 +0000 UTC Normal Pod check-span-skqwz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:59 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:59 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:01:59 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:02:13 | collector-otlp-production-grpc | 2025-01-27 08:02:12 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:02:13 | collector-otlp-production-grpc | Deleting namespace: kuttl-test-pleased-pangolin === CONT kuttl/harness/collector-autoscale logger.go:42: 08:02:20 | collector-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:02:20 | collector-autoscale | Creating namespace: kuttl-test-sharing-mallard logger.go:42: 08:02:20 | collector-autoscale/1-install | starting test step 1-install logger.go:42: 08:02:20 | collector-autoscale/1-install | Jaeger:kuttl-test-sharing-mallard/simple-prod created logger.go:42: 08:02:57 | collector-autoscale/1-install | test step completed 1-install logger.go:42: 08:02:57 | collector-autoscale/2- | starting test step 2- logger.go:42: 08:02:57 | collector-autoscale/2- | test step completed 2- logger.go:42: 08:02:57 | collector-autoscale | collector-autoscale events from ns kuttl-test-sharing-mallard: logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:28 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestsharingmallardsimpleprod-1-7d484dc58f SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestsharingmallardsimpleprod-1-7d48zjjk2 replicaset-controller logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharingmallardsimpleprod-1-7d48zjjk2 Binding Scheduled Successfully assigned kuttl-test-sharing-mallard/elasticsearch-cdm-kuttltestsharingmallardsimpleprod-1-7d48zjjk2 to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharingmallardsimpleprod-1-7d48zjjk2 AddedInterface Add eth0 [10.129.2.61/23] from ovn-kubernetes multus logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharingmallardsimpleprod-1-7d48zjjk2.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharingmallardsimpleprod-1-7d48zjjk2.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharingmallardsimpleprod-1-7d48zjjk2.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharingmallardsimpleprod-1-7d48zjjk2.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharingmallardsimpleprod-1-7d48zjjk2.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharingmallardsimpleprod-1-7d48zjjk2.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:28 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestsharingmallardsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestsharingmallardsimpleprod-1-7d484dc58f to 1 deployment-controller logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:38 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsharingmallardsimpleprod-1-7d48zjjk2.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:43 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsharingmallardsimpleprod-1-7d48zjjk2.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:54 +0000 UTC Normal Pod simple-prod-collector-cbdd5f94-h7hcf Binding Scheduled Successfully assigned kuttl-test-sharing-mallard/simple-prod-collector-cbdd5f94-h7hcf to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:54 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-cbdd5f94 SuccessfulCreate Created pod: simple-prod-collector-cbdd5f94-h7hcf replicaset-controller logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:54 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-cbdd5f94 to 1 deployment-controller logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:54 +0000 UTC Normal Pod simple-prod-query-6958944d48-c9wdt Binding Scheduled Successfully assigned kuttl-test-sharing-mallard/simple-prod-query-6958944d48-c9wdt to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:54 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6958944d48 SuccessfulCreate Created pod: simple-prod-query-6958944d48-c9wdt replicaset-controller logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:54 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6958944d48 to 1 deployment-controller logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:55 +0000 UTC Normal Pod simple-prod-collector-cbdd5f94-h7hcf AddedInterface Add eth0 [10.128.2.85/23] from ovn-kubernetes multus logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:55 +0000 UTC Normal Pod simple-prod-collector-cbdd5f94-h7hcf.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:55 +0000 UTC Normal Pod simple-prod-collector-cbdd5f94-h7hcf.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:55 +0000 UTC Normal Pod simple-prod-collector-cbdd5f94-h7hcf.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:55 +0000 UTC Normal Pod simple-prod-query-6958944d48-c9wdt AddedInterface Add eth0 [10.131.0.78/23] from ovn-kubernetes multus logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:55 +0000 UTC Normal Pod simple-prod-query-6958944d48-c9wdt.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:55 +0000 UTC Normal Pod simple-prod-query-6958944d48-c9wdt.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:55 +0000 UTC Normal Pod simple-prod-query-6958944d48-c9wdt.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:55 +0000 UTC Normal Pod simple-prod-query-6958944d48-c9wdt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:55 +0000 UTC Normal Pod simple-prod-query-6958944d48-c9wdt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:55 +0000 UTC Normal Pod simple-prod-query-6958944d48-c9wdt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:55 +0000 UTC Normal Pod simple-prod-query-6958944d48-c9wdt.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:55 +0000 UTC Normal Pod simple-prod-query-6958944d48-c9wdt.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:02:57 | collector-autoscale | 2025-01-27 08:02:55 +0000 UTC Normal Pod simple-prod-query-6958944d48-c9wdt.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:02:57 | collector-autoscale | Deleting namespace: kuttl-test-sharing-mallard === CONT kuttl/harness/collector-otlp-allinone-grpc logger.go:42: 08:03:03 | collector-otlp-allinone-grpc | Creating namespace: kuttl-test-loyal-skink logger.go:42: 08:03:03 | collector-otlp-allinone-grpc/0-install | starting test step 0-install logger.go:42: 08:03:03 | collector-otlp-allinone-grpc/0-install | Jaeger:kuttl-test-loyal-skink/my-jaeger created logger.go:42: 08:03:10 | collector-otlp-allinone-grpc/0-install | test step completed 0-install logger.go:42: 08:03:10 | collector-otlp-allinone-grpc/1-smoke-test | starting test step 1-smoke-test logger.go:42: 08:03:10 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 08:03:11 | collector-otlp-allinone-grpc/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:03:17 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 08:03:18 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:03:18 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/report-span created logger.go:42: 08:03:18 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/check-span created logger.go:42: 08:03:38 | collector-otlp-allinone-grpc/1-smoke-test | test step completed 1-smoke-test logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | collector-otlp-allinone-grpc events from ns kuttl-test-loyal-skink: logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:07 +0000 UTC Normal Pod my-jaeger-6fcf7844bc-glq9f Binding Scheduled Successfully assigned kuttl-test-loyal-skink/my-jaeger-6fcf7844bc-glq9f to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:07 +0000 UTC Normal Pod my-jaeger-6fcf7844bc-glq9f AddedInterface Add eth0 [10.128.2.86/23] from ovn-kubernetes multus logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:07 +0000 UTC Normal Pod my-jaeger-6fcf7844bc-glq9f.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:07 +0000 UTC Normal ReplicaSet.apps my-jaeger-6fcf7844bc SuccessfulCreate Created pod: my-jaeger-6fcf7844bc-glq9f replicaset-controller logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:07 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6fcf7844bc to 1 deployment-controller logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:08 +0000 UTC Normal Pod my-jaeger-6fcf7844bc-glq9f.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:08 +0000 UTC Normal Pod my-jaeger-6fcf7844bc-glq9f.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:08 +0000 UTC Normal Pod my-jaeger-6fcf7844bc-glq9f.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:08 +0000 UTC Normal Pod my-jaeger-6fcf7844bc-glq9f.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:08 +0000 UTC Normal Pod my-jaeger-6fcf7844bc-glq9f.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:14 +0000 UTC Normal Pod my-jaeger-589fdcc9b6-l7l4f Binding Scheduled Successfully assigned kuttl-test-loyal-skink/my-jaeger-589fdcc9b6-l7l4f to ip-10-0-64-118.us-east-2.compute.internal default-scheduler logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:14 +0000 UTC Normal ReplicaSet.apps my-jaeger-589fdcc9b6 SuccessfulCreate Created pod: my-jaeger-589fdcc9b6-l7l4f replicaset-controller logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:14 +0000 UTC Normal Pod my-jaeger-6fcf7844bc-glq9f.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:14 +0000 UTC Normal Pod my-jaeger-6fcf7844bc-glq9f.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:14 +0000 UTC Normal ReplicaSet.apps my-jaeger-6fcf7844bc SuccessfulDelete Deleted pod: my-jaeger-6fcf7844bc-glq9f replicaset-controller logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:14 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-6fcf7844bc to 0 from 1 deployment-controller logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:14 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-589fdcc9b6 to 1 deployment-controller logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:15 +0000 UTC Normal Pod my-jaeger-589fdcc9b6-l7l4f AddedInterface Add eth0 [10.128.2.87/23] from ovn-kubernetes multus logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:15 +0000 UTC Normal Pod my-jaeger-589fdcc9b6-l7l4f.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:15 +0000 UTC Normal Pod my-jaeger-589fdcc9b6-l7l4f.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:15 +0000 UTC Normal Pod my-jaeger-589fdcc9b6-l7l4f.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:15 +0000 UTC Normal Pod my-jaeger-589fdcc9b6-l7l4f.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:15 +0000 UTC Normal Pod my-jaeger-589fdcc9b6-l7l4f.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:15 +0000 UTC Normal Pod my-jaeger-589fdcc9b6-l7l4f.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:18 +0000 UTC Normal Pod check-span-bz2sl Binding Scheduled Successfully assigned kuttl-test-loyal-skink/check-span-bz2sl to ip-10-0-34-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:18 +0000 UTC Normal Pod check-span-bz2sl AddedInterface Add eth0 [10.131.0.79/23] from ovn-kubernetes multus logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:18 +0000 UTC Normal Pod check-span-bz2sl.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:18 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-bz2sl job-controller logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:18 +0000 UTC Normal Pod report-span-4h8bj Binding Scheduled Successfully assigned kuttl-test-loyal-skink/report-span-4h8bj to ip-10-0-70-150.us-east-2.compute.internal default-scheduler logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:18 +0000 UTC Normal Pod report-span-4h8bj AddedInterface Add eth0 [10.129.2.62/23] from ovn-kubernetes multus logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:18 +0000 UTC Normal Pod report-span-4h8bj.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:18 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-4h8bj job-controller logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:19 +0000 UTC Normal Pod check-span-bz2sl.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 549ms (549ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:19 +0000 UTC Normal Pod check-span-bz2sl.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:19 +0000 UTC Normal Pod check-span-bz2sl.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:19 +0000 UTC Normal Pod report-span-4h8bj.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 595ms (595ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:19 +0000 UTC Normal Pod report-span-4h8bj.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:19 +0000 UTC Normal Pod report-span-4h8bj.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | 2025-01-27 08:03:37 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:03:38 | collector-otlp-allinone-grpc | Deleting namespace: kuttl-test-loyal-skink === CONT kuttl/harness/cassandra-spark logger.go:42: 08:03:50 | cassandra-spark | Ignoring 01-assert.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:03:50 | cassandra-spark | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:03:50 | cassandra-spark | Creating namespace: kuttl-test-cool-bunny logger.go:42: 08:03:50 | cassandra-spark | cassandra-spark events from ns kuttl-test-cool-bunny: logger.go:42: 08:03:50 | cassandra-spark | Deleting namespace: kuttl-test-cool-bunny === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (330.76s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.84s) --- PASS: kuttl/harness/collector-otlp-allinone-http (38.13s) --- PASS: kuttl/harness/set-custom-img (46.59s) --- PASS: kuttl/harness/collector-otlp-production-http (72.48s) --- PASS: kuttl/harness/collector-otlp-production-grpc (72.37s) --- PASS: kuttl/harness/collector-autoscale (42.47s) --- PASS: kuttl/harness/collector-otlp-allinone-grpc (47.08s) --- PASS: kuttl/harness/cassandra-spark (5.77s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name miscellaneous --report --output /logs/artifacts/miscellaneous.xml ./artifacts/kuttl-report.xml time="2025-01-27T08:03:56Z" level=debug msg="Setting a new name for the test suites" time="2025-01-27T08:03:56Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-01-27T08:03:56Z" level=debug msg="normalizing test case names" time="2025-01-27T08:03:56Z" level=debug msg="miscellaneous/artifacts -> miscellaneous_artifacts" time="2025-01-27T08:03:56Z" level=debug msg="miscellaneous/collector-otlp-allinone-http -> miscellaneous_collector_otlp_allinone_http" time="2025-01-27T08:03:56Z" level=debug msg="miscellaneous/set-custom-img -> miscellaneous_set_custom_img" time="2025-01-27T08:03:56Z" level=debug msg="miscellaneous/collector-otlp-production-http -> miscellaneous_collector_otlp_production_http" time="2025-01-27T08:03:56Z" level=debug msg="miscellaneous/collector-otlp-production-grpc -> miscellaneous_collector_otlp_production_grpc" time="2025-01-27T08:03:56Z" level=debug msg="miscellaneous/collector-autoscale -> miscellaneous_collector_autoscale" time="2025-01-27T08:03:56Z" level=debug msg="miscellaneous/collector-otlp-allinone-grpc -> miscellaneous_collector_otlp_allinone_grpc" time="2025-01-27T08:03:56Z" level=debug msg="miscellaneous/cassandra-spark -> miscellaneous_cassandra_spark" +----------------------------------------------+--------+ | NAME | RESULT | +----------------------------------------------+--------+ | miscellaneous_artifacts | passed | | miscellaneous_collector_otlp_allinone_http | passed | | miscellaneous_set_custom_img | passed | | miscellaneous_collector_otlp_production_http | passed | | miscellaneous_collector_otlp_production_grpc | passed | | miscellaneous_collector_autoscale | passed | | miscellaneous_collector_otlp_allinone_grpc | passed | | miscellaneous_cassandra_spark | passed | +----------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests'