% Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 3831 100 3831 0 0 32875 0 --:--:-- --:--:-- --:--:-- 33025 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 9315 100 9315 0 0 93164 0 --:--:-- --:--:-- --:--:-- 94090 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 2607 100 2607 0 0 17813 0 --:--:-- --:--:-- --:--:-- 17856 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 1953 100 1953 0 0 21195 0 --:--:-- --:--:-- --:--:-- 21000 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 350 100 350 0 0 2137 0 --:--:-- --:--:-- --:--:-- 2147 Installing kuttl Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/kubectl-kuttl https://github.com/kudobuilder/kuttl/releases/download/v0.15.0/kubectl-kuttl_0.15.0_linux_x86_64 KUBECONFIG file is: /tmp/kubeconfig-37736524 for suite in elasticsearch examples generate upgrade sidecar streaming ui miscellaneous; do \ make run-e2e-tests-$suite ; \ done make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh elasticsearch false true + '[' 3 -ne 3 ']' + test_suite_name=elasticsearch + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/elasticsearch.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-elasticsearch make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true \ KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ ./tests/e2e/elasticsearch/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-14-130320 True False 11m Cluster version is 4.16.0-0.nightly-2024-06-14-130320' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-14-130320 True False 11m Cluster version is 4.16.0-0.nightly-2024-06-14-130320' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/elasticsearch/render.sh ++ export SUITE_DIR=./tests/e2e/elasticsearch ++ SUITE_DIR=./tests/e2e/elasticsearch ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/elasticsearch ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + start_test es-from-aio-to-production + '[' 1 -ne 1 ']' + test_name=es-from-aio-to-production + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-from-aio-to-production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-from-aio-to-production\e[0m' Rendering files for test es-from-aio-to-production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-from-aio-to-production + cd es-from-aio-to-production + jaeger_name=my-jaeger + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 03 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=03 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./03-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch.redundancyPolicy="ZeroRedundancy"' ./03-install.yaml + render_smoke_test my-jaeger true 04 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=04 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./04-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test es-increasing-replicas + '[' 1 -ne 1 ']' + test_name=es-increasing-replicas + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-increasing-replicas' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-increasing-replicas\e[0m' Rendering files for test es-increasing-replicas + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-from-aio-to-production + '[' es-from-aio-to-production '!=' _build ']' + cd .. + mkdir -p es-increasing-replicas + cd es-increasing-replicas + jaeger_name=simple-prod + '[' true = true ']' + jaeger_deployment_mode=production_autoprovisioned + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.replicas=2 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.query.replicas=2 ./02-install.yaml + cp ./01-assert.yaml ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.replicas=2 ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .status.readyReplicas=2 ./02-assert.yaml + render_smoke_test simple-prod true 03 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=03 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./03-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + cp ./02-install.yaml ./04-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.elasticsearch.nodeCount=2 ./04-install.yaml + /tmp/jaeger-tests/bin/gomplate -f ./openshift-check-es-nodes.yaml.template -o ./05-check-es-nodes.yaml + '[' true = true ']' + skip_test es-index-cleaner-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-increasing-replicas + '[' es-increasing-replicas '!=' _build ']' + cd .. + rm -rf es-index-cleaner-upstream + warning 'es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_index_cleaner -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-index-cleaner-autoprov + '[' 1 -ne 1 ']' + test_name=es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-index-cleaner-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-index-cleaner-autoprov\e[0m' Rendering files for test es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-index-cleaner-autoprov + cd es-index-cleaner-autoprov + jaeger_name=test-es-index-cleaner-with-prefix + cronjob_name=test-es-index-cleaner-with-prefix-es-index-cleaner + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + cp ../../es-index-cleaner-upstream/04-assert.yaml ../../es-index-cleaner-upstream/README.md . + render_install_jaeger test-es-index-cleaner-with-prefix production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options.es.index-prefix=""' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.enabled=false ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.numberOfDays=0 ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.esIndexCleaner.schedule="*/1 * * * *"' ./01-install.yaml + render_report_spans test-es-index-cleaner-with-prefix true 5 00 true 02 + '[' 6 -ne 6 ']' + jaeger=test-es-index-cleaner-with-prefix + is_secured=true + number_of_spans=5 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + export JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=5 + DAYS=5 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + sed 's~enabled: false~enabled: true~gi' ./01-install.yaml + CRONJOB_NAME=test-es-index-cleaner-with-prefix-es-index-cleaner + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./04-wait-es-index-cleaner.yaml + /tmp/jaeger-tests/bin/gomplate -f ./01-install.yaml -o ./05-install.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 00 06 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=00 + test_step=06 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=test-es-index-cleaner-with-prefix-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=test-es-index-cleaner-with-prefix-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./06-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./06-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.3 ++ version_ge 5.8.3 5.4 +++ echo 5.8.3 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.3 == 5.8.3 + '[' -n '' ']' + skip_test es-index-cleaner-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-index-cleaner-autoprov + '[' es-index-cleaner-autoprov '!=' _build ']' + cd .. + rm -rf es-index-cleaner-managed + warning 'es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + start_test es-multiinstance + '[' 1 -ne 1 ']' + test_name=es-multiinstance + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-multiinstance' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-multiinstance\e[0m' Rendering files for test es-multiinstance + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-multiinstance + cd es-multiinstance + jaeger_name=instance-1 + render_install_jaeger instance-1 production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=instance-1 + JAEGER_NAME=instance-1 + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f ./03-create-second-instance.yaml.template -o 03-create-second-instance.yaml + '[' true = true ']' + skip_test es-rollover-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-rollover-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-multiinstance + '[' es-multiinstance '!=' _build ']' + cd .. + rm -rf es-rollover-upstream + warning 'es-rollover-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_rollover -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-rollover-autoprov + '[' 1 -ne 1 ']' + test_name=es-rollover-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-rollover-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-rollover-autoprov\e[0m' Rendering files for test es-rollover-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-rollover-autoprov + cd es-rollover-autoprov + cp ../../es-rollover-upstream/05-assert.yaml ../../es-rollover-upstream/05-install.yaml ../../es-rollover-upstream/README.md . + jaeger_name=my-jaeger + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_report_spans my-jaeger true 2 00 true 02 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 00 03 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=00 + test_step=03 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./03-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./03-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 01 04 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=01 + test_step=04 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=01 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./04-check-indices.yaml + JOB_NUMBER=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./04-assert.yaml + render_report_spans my-jaeger true 2 02 true 06 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=02 + ensure_reported_spans=true + test_step=06 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=02 + JOB_NUMBER=02 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./06-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./06-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 02 07 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=02 + test_step=07 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=02 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./07-check-indices.yaml + JOB_NUMBER=02 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./07-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' 03 08 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + job_number=03 + test_step=08 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=03 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./08-check-indices.yaml + JOB_NUMBER=03 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./08-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' 04 09 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + job_number=04 + test_step=09 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=04 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./09-check-indices.yaml + JOB_NUMBER=04 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./09-assert.yaml + render_report_spans my-jaeger true 2 03 true 10 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=03 + ensure_reported_spans=true + test_step=10 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=03 + JOB_NUMBER=03 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./10-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./10-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + CRONJOB_NAME=my-jaeger-es-rollover + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./11-wait-rollover.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-000002'\'',' 05 11 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-000002'\'',' + job_number=05 + test_step=11 + escape_command ''\''--name'\'', '\''jaeger-span-000002'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-000002'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-000002'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-000002'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=05 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./11-check-indices.yaml + JOB_NUMBER=05 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./11-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' 06 12 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + job_number=06 + test_step=12 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=06 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./12-check-indices.yaml + JOB_NUMBER=06 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./12-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.3 ++ version_ge 5.8.3 5.4 +++ echo 5.8.3 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.3 == 5.8.3 + '[' -n '' ']' + skip_test es-rollover-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-rollover-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-rollover-autoprov + '[' es-rollover-autoprov '!=' _build ']' + cd .. + rm -rf es-rollover-managed + warning 'es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + skip_test es-spark-dependencies 'This test is not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=es-spark-dependencies + message='This test is not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + rm -rf es-spark-dependencies + warning 'es-spark-dependencies: This test is not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-spark-dependencies: This test is not supported in OpenShift\e[0m' WAR: es-spark-dependencies: This test is not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running elasticsearch E2E tests' Running elasticsearch E2E tests + cd tests/e2e/elasticsearch/_build + set +e + KUBECONFIG=/tmp/kubeconfig-37736524 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/es-from-aio-to-production === PAUSE kuttl/harness/es-from-aio-to-production === RUN kuttl/harness/es-increasing-replicas === PAUSE kuttl/harness/es-increasing-replicas === RUN kuttl/harness/es-index-cleaner-autoprov === PAUSE kuttl/harness/es-index-cleaner-autoprov === RUN kuttl/harness/es-multiinstance === PAUSE kuttl/harness/es-multiinstance === RUN kuttl/harness/es-rollover-autoprov === PAUSE kuttl/harness/es-rollover-autoprov === RUN kuttl/harness/es-simple-prod === PAUSE kuttl/harness/es-simple-prod === CONT kuttl/harness/artifacts logger.go:42: 15:30:20 | artifacts | Creating namespace: kuttl-test-live-mammoth logger.go:42: 15:30:20 | artifacts | artifacts events from ns kuttl-test-live-mammoth: logger.go:42: 15:30:20 | artifacts | Deleting namespace: kuttl-test-live-mammoth === CONT kuttl/harness/es-multiinstance logger.go:42: 15:30:26 | es-multiinstance | Ignoring 03-create-second-instance.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 15:30:26 | es-multiinstance | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 15:30:26 | es-multiinstance | Creating namespace: kuttl-test-capital-chamois logger.go:42: 15:30:26 | es-multiinstance/0-clear-namespace | starting test step 0-clear-namespace logger.go:42: 15:30:26 | es-multiinstance/0-clear-namespace | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --ignore-not-found=true] logger.go:42: 15:30:26 | es-multiinstance/0-clear-namespace | test step completed 0-clear-namespace logger.go:42: 15:30:26 | es-multiinstance/1-install | starting test step 1-install logger.go:42: 15:30:27 | es-multiinstance/1-install | Jaeger:kuttl-test-capital-chamois/instance-1 created logger.go:42: 15:31:31 | es-multiinstance/1-install | test step completed 1-install logger.go:42: 15:31:31 | es-multiinstance/2-create-namespace | starting test step 2-create-namespace logger.go:42: 15:31:31 | es-multiinstance/2-create-namespace | running command: [sh -c kubectl create namespace jaeger-e2e-multiinstance-test] logger.go:42: 15:31:31 | es-multiinstance/2-create-namespace | namespace/jaeger-e2e-multiinstance-test created logger.go:42: 15:31:31 | es-multiinstance/2-create-namespace | test step completed 2-create-namespace logger.go:42: 15:31:31 | es-multiinstance/3-create-second-instance | starting test step 3-create-second-instance logger.go:42: 15:31:31 | es-multiinstance/3-create-second-instance | running command: [sh -c kubectl apply -f ./01-install.yaml -n jaeger-e2e-multiinstance-test] logger.go:42: 15:31:34 | es-multiinstance/3-create-second-instance | jaeger.jaegertracing.io/instance-1 created logger.go:42: 15:31:34 | es-multiinstance/3-create-second-instance | running command: [sh -c /tmp/jaeger-tests/bin/kubectl-kuttl assert ./01-assert.yaml -n jaeger-e2e-multiinstance-test --timeout 1000] logger.go:42: 15:32:27 | es-multiinstance/3-create-second-instance | assert is valid logger.go:42: 15:32:27 | es-multiinstance/3-create-second-instance | test step completed 3-create-second-instance logger.go:42: 15:32:27 | es-multiinstance/4-check-secrets | starting test step 4-check-secrets logger.go:42: 15:32:27 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n $NAMESPACE > secret1] logger.go:42: 15:32:27 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n jaeger-e2e-multiinstance-test > secret2] logger.go:42: 15:32:27 | es-multiinstance/4-check-secrets | running command: [sh -c cmp --silent secret1 secret2 || exit 0] logger.go:42: 15:32:27 | es-multiinstance/4-check-secrets | test step completed 4-check-secrets logger.go:42: 15:32:27 | es-multiinstance/5-delete | starting test step 5-delete logger.go:42: 15:32:27 | es-multiinstance/5-delete | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --wait=false] logger.go:42: 15:32:27 | es-multiinstance/5-delete | namespace "jaeger-e2e-multiinstance-test" deleted logger.go:42: 15:32:27 | es-multiinstance/5-delete | test step completed 5-delete logger.go:42: 15:32:27 | es-multiinstance | es-multiinstance events from ns kuttl-test-capital-chamois: logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:30:34 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestcapitalchamoisinstance1-1-665d4758b SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestcapitalchamoisinstance1-1-665d4gx66k replicaset-controller logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:30:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcapitalchamoisinstance1-1-665d4gx66k Binding Scheduled Successfully assigned kuttl-test-capital-chamois/elasticsearch-cdm-kuttltestcapitalchamoisinstance1-1-665d4gx66k to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:30:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcapitalchamoisinstance1-1-665d4gx66k AddedInterface Add eth0 [10.129.2.20/23] from ovn-kubernetes multus logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:30:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcapitalchamoisinstance1-1-665d4gx66k.spec.containers{elasticsearch} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:30:34 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestcapitalchamoisinstance1-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestcapitalchamoisinstance1-1-665d4758b to 1 deployment-controller logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:30:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcapitalchamoisinstance1-1-665d4gx66k.spec.containers{elasticsearch} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" in 8.118s (8.118s including waiting) kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:30:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcapitalchamoisinstance1-1-665d4gx66k.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:30:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcapitalchamoisinstance1-1-665d4gx66k.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:30:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcapitalchamoisinstance1-1-665d4gx66k.spec.containers{proxy} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:30:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcapitalchamoisinstance1-1-665d4gx66k.spec.containers{proxy} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" in 4.811s (4.811s including waiting) kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:30:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcapitalchamoisinstance1-1-665d4gx66k.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:30:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcapitalchamoisinstance1-1-665d4gx66k.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:30:54 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcapitalchamoisinstance1-1-665d4gx66k.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:30:59 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcapitalchamoisinstance1-1-665d4gx66k.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:10 +0000 UTC Normal Pod instance-1-collector-6787876f8d-hdj8j Binding Scheduled Successfully assigned kuttl-test-capital-chamois/instance-1-collector-6787876f8d-hdj8j to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:10 +0000 UTC Normal ReplicaSet.apps instance-1-collector-6787876f8d SuccessfulCreate Created pod: instance-1-collector-6787876f8d-hdj8j replicaset-controller logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:10 +0000 UTC Normal Deployment.apps instance-1-collector ScalingReplicaSet Scaled up replica set instance-1-collector-6787876f8d to 1 deployment-controller logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:10 +0000 UTC Normal Pod instance-1-query-6bb779c854-hsdd9 Binding Scheduled Successfully assigned kuttl-test-capital-chamois/instance-1-query-6bb779c854-hsdd9 to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:10 +0000 UTC Normal ReplicaSet.apps instance-1-query-6bb779c854 SuccessfulCreate Created pod: instance-1-query-6bb779c854-hsdd9 replicaset-controller logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:10 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-6bb779c854 to 1 deployment-controller logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:11 +0000 UTC Warning Pod instance-1-collector-6787876f8d-hdj8j FailedMount MountVolume.SetUp failed for volume "certs" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:11 +0000 UTC Warning Pod instance-1-collector-6787876f8d-hdj8j FailedMount MountVolume.SetUp failed for volume "instance-1-collector-tls-config-volume" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:11 +0000 UTC Normal Pod instance-1-query-6bb779c854-hsdd9 AddedInterface Add eth0 [10.128.2.15/23] from ovn-kubernetes multus logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:11 +0000 UTC Normal Pod instance-1-query-6bb779c854-hsdd9.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:12 +0000 UTC Normal Pod instance-1-collector-6787876f8d-hdj8j AddedInterface Add eth0 [10.131.0.21/23] from ovn-kubernetes multus logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:12 +0000 UTC Normal Pod instance-1-collector-6787876f8d-hdj8j.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1" kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:21 +0000 UTC Normal Pod instance-1-collector-6787876f8d-hdj8j.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1" in 8.662s (8.662s including waiting) kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:21 +0000 UTC Normal Pod instance-1-collector-6787876f8d-hdj8j.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:21 +0000 UTC Normal Pod instance-1-collector-6787876f8d-hdj8j.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:21 +0000 UTC Normal Pod instance-1-query-6bb779c854-hsdd9.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" in 10.417s (10.417s including waiting) kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:21 +0000 UTC Normal Pod instance-1-query-6bb779c854-hsdd9.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:21 +0000 UTC Normal Pod instance-1-query-6bb779c854-hsdd9.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:21 +0000 UTC Normal Pod instance-1-query-6bb779c854-hsdd9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:21 +0000 UTC Normal Pod instance-1-query-6bb779c854-hsdd9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:21 +0000 UTC Normal Pod instance-1-query-6bb779c854-hsdd9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:21 +0000 UTC Normal Pod instance-1-query-6bb779c854-hsdd9.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:29 +0000 UTC Normal Pod instance-1-query-6bb779c854-hsdd9.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" in 7.412s (7.412s including waiting) kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:29 +0000 UTC Normal Pod instance-1-query-6bb779c854-hsdd9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:29 +0000 UTC Normal Pod instance-1-query-6bb779c854-hsdd9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod instance-1-collector-6787876f8d-hdj8j horizontal-pod-autoscaler logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:55 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod instance-1-collector-6787876f8d-hdj8j horizontal-pod-autoscaler logger.go:42: 15:32:27 | es-multiinstance | 2024-06-17 15:31:55 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod instance-1-collector-6787876f8d-hdj8j horizontal-pod-autoscaler logger.go:42: 15:32:27 | es-multiinstance | Deleting namespace: kuttl-test-capital-chamois === CONT kuttl/harness/es-simple-prod logger.go:42: 15:32:34 | es-simple-prod | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 15:32:34 | es-simple-prod | Creating namespace: kuttl-test-wired-barnacle logger.go:42: 15:32:34 | es-simple-prod | es-simple-prod events from ns kuttl-test-wired-barnacle: logger.go:42: 15:32:34 | es-simple-prod | Deleting namespace: kuttl-test-wired-barnacle === CONT kuttl/harness/es-rollover-autoprov logger.go:42: 15:32:41 | es-rollover-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 15:32:41 | es-rollover-autoprov | Creating namespace: kuttl-test-dominant-sculpin logger.go:42: 15:32:41 | es-rollover-autoprov/1-install | starting test step 1-install logger.go:42: 15:32:41 | es-rollover-autoprov/1-install | Jaeger:kuttl-test-dominant-sculpin/my-jaeger created logger.go:42: 15:33:16 | es-rollover-autoprov/1-install | test step completed 1-install logger.go:42: 15:33:16 | es-rollover-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 15:33:16 | es-rollover-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 15:33:20 | es-rollover-autoprov/2-report-spans | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 15:33:27 | es-rollover-autoprov/2-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 15:33:27 | es-rollover-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 15:33:28 | es-rollover-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 15:33:54 | es-rollover-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 15:33:54 | es-rollover-autoprov/3-check-indices | starting test step 3-check-indices logger.go:42: 15:33:54 | es-rollover-autoprov/3-check-indices | Job:kuttl-test-dominant-sculpin/00-check-indices created logger.go:42: 15:33:58 | es-rollover-autoprov/3-check-indices | test step completed 3-check-indices logger.go:42: 15:33:58 | es-rollover-autoprov/4-check-indices | starting test step 4-check-indices logger.go:42: 15:33:59 | es-rollover-autoprov/4-check-indices | Job:kuttl-test-dominant-sculpin/01-check-indices created logger.go:42: 15:34:03 | es-rollover-autoprov/4-check-indices | test step completed 4-check-indices logger.go:42: 15:34:03 | es-rollover-autoprov/5-install | starting test step 5-install logger.go:42: 15:34:03 | es-rollover-autoprov/5-install | Jaeger:kuttl-test-dominant-sculpin/my-jaeger updated logger.go:42: 15:34:17 | es-rollover-autoprov/5-install | test step completed 5-install logger.go:42: 15:34:17 | es-rollover-autoprov/6-report-spans | starting test step 6-report-spans logger.go:42: 15:34:17 | es-rollover-autoprov/6-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 15:34:26 | es-rollover-autoprov/6-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=02 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-02-job.yaml] logger.go:42: 15:34:27 | es-rollover-autoprov/6-report-spans | running command: [sh -c kubectl apply -f report-span-02-job.yaml -n $NAMESPACE] logger.go:42: 15:34:27 | es-rollover-autoprov/6-report-spans | job.batch/02-report-span created logger.go:42: 15:34:51 | es-rollover-autoprov/6-report-spans | test step completed 6-report-spans logger.go:42: 15:34:51 | es-rollover-autoprov/7-check-indices | starting test step 7-check-indices logger.go:42: 15:34:52 | es-rollover-autoprov/7-check-indices | Job:kuttl-test-dominant-sculpin/02-check-indices created logger.go:42: 15:34:59 | es-rollover-autoprov/7-check-indices | test step completed 7-check-indices logger.go:42: 15:34:59 | es-rollover-autoprov/8-check-indices | starting test step 8-check-indices logger.go:42: 15:34:59 | es-rollover-autoprov/8-check-indices | Job:kuttl-test-dominant-sculpin/03-check-indices created logger.go:42: 15:35:04 | es-rollover-autoprov/8-check-indices | test step completed 8-check-indices logger.go:42: 15:35:04 | es-rollover-autoprov/9-check-indices | starting test step 9-check-indices logger.go:42: 15:35:04 | es-rollover-autoprov/9-check-indices | Job:kuttl-test-dominant-sculpin/04-check-indices created logger.go:42: 15:35:09 | es-rollover-autoprov/9-check-indices | test step completed 9-check-indices logger.go:42: 15:35:09 | es-rollover-autoprov/10-report-spans | starting test step 10-report-spans logger.go:42: 15:35:09 | es-rollover-autoprov/10-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 15:35:17 | es-rollover-autoprov/10-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=03 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-03-job.yaml] logger.go:42: 15:35:18 | es-rollover-autoprov/10-report-spans | running command: [sh -c kubectl apply -f report-span-03-job.yaml -n $NAMESPACE] logger.go:42: 15:35:18 | es-rollover-autoprov/10-report-spans | job.batch/03-report-span created logger.go:42: 15:35:41 | es-rollover-autoprov/10-report-spans | test step completed 10-report-spans logger.go:42: 15:35:41 | es-rollover-autoprov/11-check-indices | starting test step 11-check-indices logger.go:42: 15:35:41 | es-rollover-autoprov/11-check-indices | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob my-jaeger-es-rollover --namespace $NAMESPACE] logger.go:42: 15:35:51 | es-rollover-autoprov/11-check-indices | time="2024-06-17T15:35:51Z" level=debug msg="Checking if the my-jaeger-es-rollover CronJob exists" logger.go:42: 15:35:51 | es-rollover-autoprov/11-check-indices | time="2024-06-17T15:35:51Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 15:35:51 | es-rollover-autoprov/11-check-indices | time="2024-06-17T15:35:51Z" level=info msg="Cronjob my-jaeger-es-rollover found successfully" logger.go:42: 15:35:51 | es-rollover-autoprov/11-check-indices | time="2024-06-17T15:35:51Z" level=debug msg="Waiting for the next scheduled job from my-jaeger-es-rollover cronjob" logger.go:42: 15:35:51 | es-rollover-autoprov/11-check-indices | time="2024-06-17T15:35:51Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 15:36:01 | es-rollover-autoprov/11-check-indices | time="2024-06-17T15:36:01Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 15:36:11 | es-rollover-autoprov/11-check-indices | time="2024-06-17T15:36:11Z" level=info msg="Job of owner my-jaeger-es-rollover succeeded after my-jaeger-es-rollover 20.059021273s" logger.go:42: 15:36:12 | es-rollover-autoprov/11-check-indices | Job:kuttl-test-dominant-sculpin/05-check-indices created logger.go:42: 15:36:16 | es-rollover-autoprov/11-check-indices | test step completed 11-check-indices logger.go:42: 15:36:16 | es-rollover-autoprov/12-check-indices | starting test step 12-check-indices logger.go:42: 15:36:16 | es-rollover-autoprov/12-check-indices | Job:kuttl-test-dominant-sculpin/06-check-indices created logger.go:42: 15:36:20 | es-rollover-autoprov/12-check-indices | test step completed 12-check-indices logger.go:42: 15:36:20 | es-rollover-autoprov | es-rollover-autoprov events from ns kuttl-test-dominant-sculpin: logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:32:46 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestdominantsculpinmyjaeger-1-694c4bb666 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestdominantsculpinmyjaeger-1-694c4hzvnd replicaset-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:32:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdominantsculpinmyjaeger-1-694c4hzvnd Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/elasticsearch-cdm-kuttltestdominantsculpinmyjaeger-1-694c4hzvnd to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:32:46 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestdominantsculpinmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestdominantsculpinmyjaeger-1-694c4bb666 to 1 deployment-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:32:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdominantsculpinmyjaeger-1-694c4hzvnd AddedInterface Add eth0 [10.129.2.21/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:32:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdominantsculpinmyjaeger-1-694c4hzvnd.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:32:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdominantsculpinmyjaeger-1-694c4hzvnd.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:32:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdominantsculpinmyjaeger-1-694c4hzvnd.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:32:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdominantsculpinmyjaeger-1-694c4hzvnd.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:32:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdominantsculpinmyjaeger-1-694c4hzvnd.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:32:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdominantsculpinmyjaeger-1-694c4hzvnd.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:32:57 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestdominantsculpinmyjaeger-1-694c4hzvnd.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:02 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestdominantsculpinmyjaeger-1-694c4hzvnd.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:13 +0000 UTC Normal Pod my-jaeger-collector-85f66554fb-g99nx Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/my-jaeger-collector-85f66554fb-g99nx to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:13 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-85f66554fb SuccessfulCreate Created pod: my-jaeger-collector-85f66554fb-g99nx replicaset-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:13 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-85f66554fb to 1 deployment-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:13 +0000 UTC Normal Pod my-jaeger-query-7bdb8f76dc-d5zwz Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/my-jaeger-query-7bdb8f76dc-d5zwz to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:13 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7bdb8f76dc SuccessfulCreate Created pod: my-jaeger-query-7bdb8f76dc-d5zwz replicaset-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:13 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7bdb8f76dc to 1 deployment-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:14 +0000 UTC Normal Pod my-jaeger-collector-85f66554fb-g99nx AddedInterface Add eth0 [10.131.0.23/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:14 +0000 UTC Normal Pod my-jaeger-collector-85f66554fb-g99nx.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1" already present on machine kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:14 +0000 UTC Normal Pod my-jaeger-collector-85f66554fb-g99nx.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:14 +0000 UTC Normal Pod my-jaeger-collector-85f66554fb-g99nx.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:14 +0000 UTC Normal Pod my-jaeger-query-7bdb8f76dc-d5zwz AddedInterface Add eth0 [10.128.2.18/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:14 +0000 UTC Normal Pod my-jaeger-query-7bdb8f76dc-d5zwz.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:14 +0000 UTC Normal Pod my-jaeger-query-7bdb8f76dc-d5zwz.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:14 +0000 UTC Normal Pod my-jaeger-query-7bdb8f76dc-d5zwz.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:14 +0000 UTC Normal Pod my-jaeger-query-7bdb8f76dc-d5zwz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:14 +0000 UTC Normal Pod my-jaeger-query-7bdb8f76dc-d5zwz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:14 +0000 UTC Normal Pod my-jaeger-query-7bdb8f76dc-d5zwz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:14 +0000 UTC Normal Pod my-jaeger-query-7bdb8f76dc-d5zwz.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:14 +0000 UTC Normal Pod my-jaeger-query-7bdb8f76dc-d5zwz.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:14 +0000 UTC Normal Pod my-jaeger-query-7bdb8f76dc-d5zwz.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:21 +0000 UTC Normal Pod my-jaeger-query-7bdb8f76dc-d5zwz.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:21 +0000 UTC Normal Pod my-jaeger-query-7bdb8f76dc-d5zwz.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:21 +0000 UTC Normal Pod my-jaeger-query-7bdb8f76dc-d5zwz.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:21 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7bdb8f76dc SuccessfulDelete Deleted pod: my-jaeger-query-7bdb8f76dc-d5zwz replicaset-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:21 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-7bdb8f76dc to 0 from 1 deployment-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:22 +0000 UTC Normal Pod my-jaeger-query-67b89fdcd-4wl7g Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/my-jaeger-query-67b89fdcd-4wl7g to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:22 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-67b89fdcd SuccessfulCreate Created pod: my-jaeger-query-67b89fdcd-4wl7g replicaset-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:22 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-67b89fdcd to 1 deployment-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:23 +0000 UTC Normal Pod my-jaeger-query-67b89fdcd-4wl7g AddedInterface Add eth0 [10.128.2.19/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:23 +0000 UTC Normal Pod my-jaeger-query-67b89fdcd-4wl7g.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:23 +0000 UTC Normal Pod my-jaeger-query-67b89fdcd-4wl7g.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:23 +0000 UTC Normal Pod my-jaeger-query-67b89fdcd-4wl7g.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:23 +0000 UTC Normal Pod my-jaeger-query-67b89fdcd-4wl7g.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:23 +0000 UTC Normal Pod my-jaeger-query-67b89fdcd-4wl7g.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:23 +0000 UTC Normal Pod my-jaeger-query-67b89fdcd-4wl7g.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:23 +0000 UTC Normal Pod my-jaeger-query-67b89fdcd-4wl7g.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:23 +0000 UTC Normal Pod my-jaeger-query-67b89fdcd-4wl7g.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:23 +0000 UTC Normal Pod my-jaeger-query-67b89fdcd-4wl7g.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:28 +0000 UTC Normal Pod 00-report-span-2q9jf Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/00-report-span-2q9jf to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:28 +0000 UTC Normal Pod 00-report-span-2q9jf AddedInterface Add eth0 [10.131.0.24/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:28 +0000 UTC Normal Pod 00-report-span-2q9jf.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:28 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-2q9jf job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:28 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:28 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:28 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:31 +0000 UTC Normal Pod 00-report-span-2q9jf.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.614s (2.614s including waiting) kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:31 +0000 UTC Normal Pod 00-report-span-2q9jf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:31 +0000 UTC Normal Pod 00-report-span-2q9jf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-85f66554fb-g99nx horizontal-pod-autoscaler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:54 +0000 UTC Normal Pod 00-check-indices-lxtvd Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/00-check-indices-lxtvd to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:54 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-lxtvd job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:54 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:55 +0000 UTC Normal Pod 00-check-indices-lxtvd AddedInterface Add eth0 [10.131.0.25/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:55 +0000 UTC Normal Pod 00-check-indices-lxtvd.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:55 +0000 UTC Normal Pod 00-check-indices-lxtvd.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 323ms (323ms including waiting) kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:55 +0000 UTC Normal Pod 00-check-indices-lxtvd.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:55 +0000 UTC Normal Pod 00-check-indices-lxtvd.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:58 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-85f66554fb-g99nx horizontal-pod-autoscaler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-85f66554fb-g99nx horizontal-pod-autoscaler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:59 +0000 UTC Normal Pod 01-check-indices-x467m Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/01-check-indices-x467m to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:59 +0000 UTC Normal Pod 01-check-indices-x467m AddedInterface Add eth0 [10.131.0.26/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:59 +0000 UTC Normal Pod 01-check-indices-x467m.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:59 +0000 UTC Normal Pod 01-check-indices-x467m.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 303ms (303ms including waiting) kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:33:59 +0000 UTC Normal Job.batch 01-check-indices SuccessfulCreate Created pod: 01-check-indices-x467m job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:00 +0000 UTC Normal Pod 01-check-indices-x467m.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:00 +0000 UTC Normal Pod 01-check-indices-x467m.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:02 +0000 UTC Normal Job.batch 01-check-indices Completed Job completed job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:04 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-pvhh8 Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/my-jaeger-es-rollover-create-mapping-pvhh8 to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:04 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping SuccessfulCreate Created pod: my-jaeger-es-rollover-create-mapping-pvhh8 job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:05 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-pvhh8 AddedInterface Add eth0 [10.131.0.27/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:05 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-pvhh8.spec.containers{my-jaeger-es-rollover-create-mapping} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:1ddb96f45754b0106ab6f841c95791e2310128852bd7d0e74051a3b48443f97d" kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:13 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-pvhh8.spec.containers{my-jaeger-es-rollover-create-mapping} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:1ddb96f45754b0106ab6f841c95791e2310128852bd7d0e74051a3b48443f97d" in 8.269s (8.269s including waiting) kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:13 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-pvhh8.spec.containers{my-jaeger-es-rollover-create-mapping} Created Created container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:13 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-pvhh8.spec.containers{my-jaeger-es-rollover-create-mapping} Started Started container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:16 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping Completed Job completed job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:17 +0000 UTC Normal Pod my-jaeger-collector-85f66554fb-g99nx.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:17 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-85f66554fb SuccessfulDelete Deleted pod: my-jaeger-collector-85f66554fb-g99nx replicaset-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:17 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-85f66554fb to 0 from 1 deployment-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:17 +0000 UTC Normal Pod my-jaeger-query-67b89fdcd-4wl7g.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:17 +0000 UTC Normal Pod my-jaeger-query-67b89fdcd-4wl7g.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:17 +0000 UTC Normal Pod my-jaeger-query-67b89fdcd-4wl7g.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:17 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-67b89fdcd SuccessfulDelete Deleted pod: my-jaeger-query-67b89fdcd-4wl7g replicaset-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:17 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-67b89fdcd to 0 from 1 deployment-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:18 +0000 UTC Normal Pod my-jaeger-collector-58dcc8c4f-hrq8f Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/my-jaeger-collector-58dcc8c4f-hrq8f to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:18 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-58dcc8c4f SuccessfulCreate Created pod: my-jaeger-collector-58dcc8c4f-hrq8f replicaset-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:18 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-58dcc8c4f to 1 deployment-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:18 +0000 UTC Normal Pod my-jaeger-query-65f8854889-wtrfb Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/my-jaeger-query-65f8854889-wtrfb to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:18 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-65f8854889 SuccessfulCreate Created pod: my-jaeger-query-65f8854889-wtrfb replicaset-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:18 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-65f8854889 to 1 deployment-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:19 +0000 UTC Normal Pod my-jaeger-collector-58dcc8c4f-hrq8f AddedInterface Add eth0 [10.131.0.28/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:19 +0000 UTC Normal Pod my-jaeger-collector-58dcc8c4f-hrq8f.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1" already present on machine kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:19 +0000 UTC Normal Pod my-jaeger-collector-58dcc8c4f-hrq8f.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:19 +0000 UTC Normal Pod my-jaeger-collector-58dcc8c4f-hrq8f.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:19 +0000 UTC Normal Pod my-jaeger-query-65f8854889-wtrfb AddedInterface Add eth0 [10.128.2.20/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:19 +0000 UTC Normal Pod my-jaeger-query-65f8854889-wtrfb.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:19 +0000 UTC Normal Pod my-jaeger-query-65f8854889-wtrfb.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:19 +0000 UTC Normal Pod my-jaeger-query-65f8854889-wtrfb.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:19 +0000 UTC Normal Pod my-jaeger-query-65f8854889-wtrfb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:19 +0000 UTC Normal Pod my-jaeger-query-65f8854889-wtrfb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:19 +0000 UTC Normal Pod my-jaeger-query-65f8854889-wtrfb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:19 +0000 UTC Normal Pod my-jaeger-query-65f8854889-wtrfb.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:19 +0000 UTC Normal Pod my-jaeger-query-65f8854889-wtrfb.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:19 +0000 UTC Normal Pod my-jaeger-query-65f8854889-wtrfb.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:27 +0000 UTC Normal Pod 02-report-span-x7b5f Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/02-report-span-x7b5f to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:27 +0000 UTC Normal Job.batch 02-report-span SuccessfulCreate Created pod: 02-report-span-x7b5f job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:28 +0000 UTC Normal Pod 02-report-span-x7b5f AddedInterface Add eth0 [10.131.0.29/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:28 +0000 UTC Normal Pod 02-report-span-x7b5f.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:28 +0000 UTC Normal Pod 02-report-span-x7b5f.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 354ms (354ms including waiting) kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:28 +0000 UTC Normal Pod 02-report-span-x7b5f.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:28 +0000 UTC Normal Pod 02-report-span-x7b5f.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:51 +0000 UTC Normal Job.batch 02-report-span Completed Job completed job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:52 +0000 UTC Normal Pod 02-check-indices-hngb5 Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/02-check-indices-hngb5 to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:52 +0000 UTC Normal Job.batch 02-check-indices SuccessfulCreate Created pod: 02-check-indices-hngb5 job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:53 +0000 UTC Normal Pod 02-check-indices-hngb5 AddedInterface Add eth0 [10.131.0.30/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:53 +0000 UTC Normal Pod 02-check-indices-hngb5.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:55 +0000 UTC Normal Pod 02-check-indices-hngb5.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.588s (2.588s including waiting) kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:55 +0000 UTC Normal Pod 02-check-indices-hngb5.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:55 +0000 UTC Normal Pod 02-check-indices-hngb5.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:58 +0000 UTC Normal Job.batch 02-check-indices Completed Job completed job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-58dcc8c4f-hrq8f horizontal-pod-autoscaler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-58dcc8c4f-hrq8f horizontal-pod-autoscaler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-58dcc8c4f-hrq8f horizontal-pod-autoscaler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:59 +0000 UTC Normal Pod 03-check-indices-ptxfw Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/03-check-indices-ptxfw to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:34:59 +0000 UTC Normal Job.batch 03-check-indices SuccessfulCreate Created pod: 03-check-indices-ptxfw job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal Pod 03-check-indices-ptxfw AddedInterface Add eth0 [10.131.0.31/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal Pod 03-check-indices-ptxfw.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal Pod 03-check-indices-ptxfw.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 381ms (381ms including waiting) kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal Pod 03-check-indices-ptxfw.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal Pod 03-check-indices-ptxfw.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28643975-cx6tg Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/my-jaeger-es-lookback-28643975-cx6tg to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28643975-cx6tg AddedInterface Add eth0 [10.131.0.33/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28643975-cx6tg.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:1ddb96f45754b0106ab6f841c95791e2310128852bd7d0e74051a3b48443f97d" already present on machine kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28643975-cx6tg.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28643975-cx6tg.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28643975 SuccessfulCreate Created pod: my-jaeger-es-lookback-28643975-cx6tg job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28643975 cronjob-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28643975-hcxbj Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/my-jaeger-es-rollover-28643975-hcxbj to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28643975-hcxbj AddedInterface Add eth0 [10.131.0.32/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28643975-hcxbj.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:1ddb96f45754b0106ab6f841c95791e2310128852bd7d0e74051a3b48443f97d" already present on machine kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28643975-hcxbj.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28643975-hcxbj.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28643975 SuccessfulCreate Created pod: my-jaeger-es-rollover-28643975-hcxbj job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28643975 cronjob-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:03 +0000 UTC Normal Job.batch 03-check-indices Completed Job completed job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28643975 Completed Job completed job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28643975, status: Complete cronjob-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28643975 Completed Job completed job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28643975, status: Complete cronjob-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:04 +0000 UTC Normal Pod 04-check-indices-l9vwc Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/04-check-indices-l9vwc to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:04 +0000 UTC Normal Job.batch 04-check-indices SuccessfulCreate Created pod: 04-check-indices-l9vwc job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:05 +0000 UTC Normal Pod 04-check-indices-l9vwc AddedInterface Add eth0 [10.131.0.34/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:05 +0000 UTC Normal Pod 04-check-indices-l9vwc.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:05 +0000 UTC Normal Pod 04-check-indices-l9vwc.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 351ms (351ms including waiting) kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:05 +0000 UTC Normal Pod 04-check-indices-l9vwc.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:05 +0000 UTC Normal Pod 04-check-indices-l9vwc.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:08 +0000 UTC Normal Job.batch 04-check-indices Completed Job completed job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:18 +0000 UTC Normal Pod 03-report-span-q9q6n Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/03-report-span-q9q6n to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:18 +0000 UTC Normal Job.batch 03-report-span SuccessfulCreate Created pod: 03-report-span-q9q6n job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:19 +0000 UTC Normal Pod 03-report-span-q9q6n AddedInterface Add eth0 [10.131.0.35/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:19 +0000 UTC Normal Pod 03-report-span-q9q6n.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:19 +0000 UTC Normal Pod 03-report-span-q9q6n.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 331ms (331ms including waiting) kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:19 +0000 UTC Normal Pod 03-report-span-q9q6n.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:19 +0000 UTC Normal Pod 03-report-span-q9q6n.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:35:41 +0000 UTC Normal Job.batch 03-report-span Completed Job completed job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28643976-6pn6z Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/my-jaeger-es-lookback-28643976-6pn6z to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28643976 SuccessfulCreate Created pod: my-jaeger-es-lookback-28643976-6pn6z job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28643976 cronjob-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28643976-r98tf Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/my-jaeger-es-rollover-28643976-r98tf to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28643976 SuccessfulCreate Created pod: my-jaeger-es-rollover-28643976-r98tf job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28643976 cronjob-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28643976-6pn6z AddedInterface Add eth0 [10.131.0.36/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28643976-6pn6z.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:1ddb96f45754b0106ab6f841c95791e2310128852bd7d0e74051a3b48443f97d" already present on machine kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28643976-6pn6z.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28643976-6pn6z.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:01 +0000 UTC Normal Pod my-jaeger-es-rollover-28643976-r98tf AddedInterface Add eth0 [10.131.0.37/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:01 +0000 UTC Normal Pod my-jaeger-es-rollover-28643976-r98tf.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:1ddb96f45754b0106ab6f841c95791e2310128852bd7d0e74051a3b48443f97d" already present on machine kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:01 +0000 UTC Normal Pod my-jaeger-es-rollover-28643976-r98tf.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:01 +0000 UTC Normal Pod my-jaeger-es-rollover-28643976-r98tf.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28643976 Completed Job completed job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28643976, status: Complete cronjob-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28643976 Completed Job completed job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28643976, status: Complete cronjob-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:12 +0000 UTC Normal Pod 05-check-indices-2tx9w Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/05-check-indices-2tx9w to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:12 +0000 UTC Normal Pod 05-check-indices-2tx9w AddedInterface Add eth0 [10.131.0.38/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:12 +0000 UTC Normal Pod 05-check-indices-2tx9w.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:12 +0000 UTC Normal Job.batch 05-check-indices SuccessfulCreate Created pod: 05-check-indices-2tx9w job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:13 +0000 UTC Normal Pod 05-check-indices-2tx9w.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 382ms (382ms including waiting) kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:13 +0000 UTC Normal Pod 05-check-indices-2tx9w.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:13 +0000 UTC Normal Pod 05-check-indices-2tx9w.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:15 +0000 UTC Normal Job.batch 05-check-indices Completed Job completed job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:16 +0000 UTC Normal Pod 06-check-indices-dn6ss Binding Scheduled Successfully assigned kuttl-test-dominant-sculpin/06-check-indices-dn6ss to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:16 +0000 UTC Normal Pod 06-check-indices-dn6ss AddedInterface Add eth0 [10.131.0.39/23] from ovn-kubernetes multus logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:16 +0000 UTC Normal Pod 06-check-indices-dn6ss.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:16 +0000 UTC Normal Job.batch 06-check-indices SuccessfulCreate Created pod: 06-check-indices-dn6ss job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:17 +0000 UTC Normal Pod 06-check-indices-dn6ss.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 375ms (375ms including waiting) kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:17 +0000 UTC Normal Pod 06-check-indices-dn6ss.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:17 +0000 UTC Normal Pod 06-check-indices-dn6ss.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 15:36:20 | es-rollover-autoprov | 2024-06-17 15:36:19 +0000 UTC Normal Job.batch 06-check-indices Completed Job completed job-controller logger.go:42: 15:36:20 | es-rollover-autoprov | Deleting namespace: kuttl-test-dominant-sculpin === CONT kuttl/harness/es-increasing-replicas logger.go:42: 15:36:28 | es-increasing-replicas | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 15:36:28 | es-increasing-replicas | Ignoring check-es-nodes.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 15:36:28 | es-increasing-replicas | Ignoring openshift-check-es-nodes.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 15:36:28 | es-increasing-replicas | Creating namespace: kuttl-test-teaching-tick logger.go:42: 15:36:28 | es-increasing-replicas/1-install | starting test step 1-install logger.go:42: 15:36:28 | es-increasing-replicas/1-install | Jaeger:kuttl-test-teaching-tick/simple-prod created logger.go:42: 15:37:04 | es-increasing-replicas/1-install | test step completed 1-install logger.go:42: 15:37:04 | es-increasing-replicas/2-install | starting test step 2-install logger.go:42: 15:37:05 | es-increasing-replicas/2-install | Jaeger:kuttl-test-teaching-tick/simple-prod updated logger.go:42: 15:37:14 | es-increasing-replicas/2-install | test step completed 2-install logger.go:42: 15:37:14 | es-increasing-replicas/3-smoke-test | starting test step 3-smoke-test logger.go:42: 15:37:14 | es-increasing-replicas/3-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 15:37:15 | es-increasing-replicas/3-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 15:37:22 | es-increasing-replicas/3-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 15:37:22 | es-increasing-replicas/3-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 15:37:23 | es-increasing-replicas/3-smoke-test | job.batch/report-span created logger.go:42: 15:37:23 | es-increasing-replicas/3-smoke-test | job.batch/check-span created logger.go:42: 15:37:36 | es-increasing-replicas/3-smoke-test | test step completed 3-smoke-test logger.go:42: 15:37:36 | es-increasing-replicas/4-install | starting test step 4-install logger.go:42: 15:37:36 | es-increasing-replicas/4-install | Jaeger:kuttl-test-teaching-tick/simple-prod updated logger.go:42: 15:37:36 | es-increasing-replicas/4-install | test step completed 4-install logger.go:42: 15:37:36 | es-increasing-replicas/5-check-es-nodes | starting test step 5-check-es-nodes logger.go:42: 15:37:36 | es-increasing-replicas/5-check-es-nodes | running command: [sh -c ./check-es-nodes.sh $NAMESPACE] logger.go:42: 15:37:36 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 15:37:37 | es-increasing-replicas/5-check-es-nodes | false logger.go:42: 15:37:37 | es-increasing-replicas/5-check-es-nodes | Error: no matches found logger.go:42: 15:37:42 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 15:37:42 | es-increasing-replicas/5-check-es-nodes | true logger.go:42: 15:37:42 | es-increasing-replicas/5-check-es-nodes | test step completed 5-check-es-nodes logger.go:42: 15:37:42 | es-increasing-replicas | es-increasing-replicas events from ns kuttl-test-teaching-tick: logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:36:34 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestteachingticksimpleprod-1-66f4c8d57d SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestteachingticksimpleprod-1-66f4c8pjxx2 replicaset-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:36:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingticksimpleprod-1-66f4c8pjxx2 Binding Scheduled Successfully assigned kuttl-test-teaching-tick/elasticsearch-cdm-kuttltestteachingticksimpleprod-1-66f4c8pjxx2 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:36:34 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestteachingticksimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestteachingticksimpleprod-1-66f4c8d57d to 1 deployment-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:36:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingticksimpleprod-1-66f4c8pjxx2 AddedInterface Add eth0 [10.129.2.22/23] from ovn-kubernetes multus logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:36:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingticksimpleprod-1-66f4c8pjxx2.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:36:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingticksimpleprod-1-66f4c8pjxx2.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:36:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingticksimpleprod-1-66f4c8pjxx2.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:36:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingticksimpleprod-1-66f4c8pjxx2.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:36:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingticksimpleprod-1-66f4c8pjxx2.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:36:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingticksimpleprod-1-66f4c8pjxx2.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:36:45 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestteachingticksimpleprod-1-66f4c8pjxx2.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:36:50 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestteachingticksimpleprod-1-66f4c8pjxx2.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:01 +0000 UTC Normal Pod simple-prod-collector-5cd647b47b-5hzc7 Binding Scheduled Successfully assigned kuttl-test-teaching-tick/simple-prod-collector-5cd647b47b-5hzc7 to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:01 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5cd647b47b SuccessfulCreate Created pod: simple-prod-collector-5cd647b47b-5hzc7 replicaset-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:01 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-5cd647b47b to 1 deployment-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:01 +0000 UTC Normal Pod simple-prod-query-7dd7876945-rrcpd Binding Scheduled Successfully assigned kuttl-test-teaching-tick/simple-prod-query-7dd7876945-rrcpd to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:01 +0000 UTC Normal ReplicaSet.apps simple-prod-query-7dd7876945 SuccessfulCreate Created pod: simple-prod-query-7dd7876945-rrcpd replicaset-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:01 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-7dd7876945 to 1 deployment-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:02 +0000 UTC Normal Pod simple-prod-collector-5cd647b47b-5hzc7 AddedInterface Add eth0 [10.131.0.40/23] from ovn-kubernetes multus logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:02 +0000 UTC Normal Pod simple-prod-collector-5cd647b47b-5hzc7.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:02 +0000 UTC Normal Pod simple-prod-collector-5cd647b47b-5hzc7.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:02 +0000 UTC Normal Pod simple-prod-collector-5cd647b47b-5hzc7.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:02 +0000 UTC Normal Pod simple-prod-query-7dd7876945-rrcpd AddedInterface Add eth0 [10.128.2.21/23] from ovn-kubernetes multus logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:02 +0000 UTC Normal Pod simple-prod-query-7dd7876945-rrcpd.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:02 +0000 UTC Normal Pod simple-prod-query-7dd7876945-rrcpd.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:02 +0000 UTC Normal Pod simple-prod-query-7dd7876945-rrcpd.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:02 +0000 UTC Normal Pod simple-prod-query-7dd7876945-rrcpd.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:02 +0000 UTC Normal Pod simple-prod-query-7dd7876945-rrcpd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:02 +0000 UTC Normal Pod simple-prod-query-7dd7876945-rrcpd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:02 +0000 UTC Normal Pod simple-prod-query-7dd7876945-rrcpd.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:02 +0000 UTC Normal Pod simple-prod-query-7dd7876945-rrcpd.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:02 +0000 UTC Normal Pod simple-prod-query-7dd7876945-rrcpd.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:06 +0000 UTC Normal Pod simple-prod-collector-5cd647b47b-r6chb Binding Scheduled Successfully assigned kuttl-test-teaching-tick/simple-prod-collector-5cd647b47b-r6chb to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:06 +0000 UTC Normal Pod simple-prod-collector-5cd647b47b-r6chb AddedInterface Add eth0 [10.128.2.22/23] from ovn-kubernetes multus logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:06 +0000 UTC Normal Pod simple-prod-collector-5cd647b47b-r6chb.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:06 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5cd647b47b SuccessfulCreate Created pod: simple-prod-collector-5cd647b47b-r6chb replicaset-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:06 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-5cd647b47b to 2 from 1 deployment-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:06 +0000 UTC Normal Pod simple-prod-query-7dd7876945-qtpd9 Binding Scheduled Successfully assigned kuttl-test-teaching-tick/simple-prod-query-7dd7876945-qtpd9 to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:06 +0000 UTC Normal Pod simple-prod-query-7dd7876945-qtpd9 AddedInterface Add eth0 [10.131.0.41/23] from ovn-kubernetes multus logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:06 +0000 UTC Normal Pod simple-prod-query-7dd7876945-qtpd9.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:06 +0000 UTC Normal ReplicaSet.apps simple-prod-query-7dd7876945 SuccessfulCreate Created pod: simple-prod-query-7dd7876945-qtpd9 replicaset-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:06 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-7dd7876945 to 2 from 1 deployment-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:07 +0000 UTC Normal Pod simple-prod-collector-5cd647b47b-r6chb.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:07 +0000 UTC Normal Pod simple-prod-collector-5cd647b47b-r6chb.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:10 +0000 UTC Normal Pod simple-prod-query-7dd7876945-qtpd9.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" in 3.301s (3.301s including waiting) kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:10 +0000 UTC Normal Pod simple-prod-query-7dd7876945-qtpd9.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:10 +0000 UTC Normal Pod simple-prod-query-7dd7876945-qtpd9.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:10 +0000 UTC Normal Pod simple-prod-query-7dd7876945-qtpd9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:10 +0000 UTC Normal Pod simple-prod-query-7dd7876945-qtpd9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:10 +0000 UTC Normal Pod simple-prod-query-7dd7876945-qtpd9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:10 +0000 UTC Normal Pod simple-prod-query-7dd7876945-qtpd9.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:12 +0000 UTC Normal Pod simple-prod-query-7dd7876945-qtpd9.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" in 2.206s (2.206s including waiting) kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:12 +0000 UTC Normal Pod simple-prod-query-7dd7876945-qtpd9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:12 +0000 UTC Normal Pod simple-prod-query-7dd7876945-qtpd9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:17 +0000 UTC Normal Pod simple-prod-query-7dd7876945-qtpd9.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:17 +0000 UTC Normal Pod simple-prod-query-7dd7876945-qtpd9.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:17 +0000 UTC Normal Pod simple-prod-query-7dd7876945-qtpd9.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:17 +0000 UTC Normal Pod simple-prod-query-7dd7876945-rrcpd.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:17 +0000 UTC Normal Pod simple-prod-query-7dd7876945-rrcpd.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:17 +0000 UTC Normal Pod simple-prod-query-7dd7876945-rrcpd.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:17 +0000 UTC Normal ReplicaSet.apps simple-prod-query-7dd7876945 SuccessfulDelete Deleted pod: simple-prod-query-7dd7876945-qtpd9 replicaset-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:17 +0000 UTC Normal ReplicaSet.apps simple-prod-query-7dd7876945 SuccessfulDelete Deleted pod: simple-prod-query-7dd7876945-rrcpd replicaset-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:17 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-7dd7876945 to 0 from 2 deployment-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:17 +0000 UTC Warning Endpoints simple-prod-query FailedToUpdateEndpoint Failed to update endpoint kuttl-test-teaching-tick/simple-prod-query: Operation cannot be fulfilled on endpoints "simple-prod-query": the object has been modified; please apply your changes to the latest version and try again endpoint-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:18 +0000 UTC Normal Pod simple-prod-query-b69966796-j6sqr Binding Scheduled Successfully assigned kuttl-test-teaching-tick/simple-prod-query-b69966796-j6sqr to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:18 +0000 UTC Normal Pod simple-prod-query-b69966796-j6sqr AddedInterface Add eth0 [10.128.2.23/23] from ovn-kubernetes multus logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:18 +0000 UTC Normal Pod simple-prod-query-b69966796-j6sqr.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:18 +0000 UTC Normal Pod simple-prod-query-b69966796-ldqdw Binding Scheduled Successfully assigned kuttl-test-teaching-tick/simple-prod-query-b69966796-ldqdw to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:18 +0000 UTC Normal Pod simple-prod-query-b69966796-ldqdw AddedInterface Add eth0 [10.131.0.42/23] from ovn-kubernetes multus logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:18 +0000 UTC Normal Pod simple-prod-query-b69966796-ldqdw.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:18 +0000 UTC Normal ReplicaSet.apps simple-prod-query-b69966796 SuccessfulCreate Created pod: simple-prod-query-b69966796-ldqdw replicaset-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:18 +0000 UTC Normal ReplicaSet.apps simple-prod-query-b69966796 SuccessfulCreate Created pod: simple-prod-query-b69966796-j6sqr replicaset-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:18 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-b69966796 to 2 deployment-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:19 +0000 UTC Normal Pod simple-prod-query-b69966796-j6sqr.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:19 +0000 UTC Normal Pod simple-prod-query-b69966796-j6sqr.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:19 +0000 UTC Normal Pod simple-prod-query-b69966796-j6sqr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:19 +0000 UTC Normal Pod simple-prod-query-b69966796-j6sqr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:19 +0000 UTC Normal Pod simple-prod-query-b69966796-j6sqr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:19 +0000 UTC Normal Pod simple-prod-query-b69966796-j6sqr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:19 +0000 UTC Normal Pod simple-prod-query-b69966796-j6sqr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:19 +0000 UTC Normal Pod simple-prod-query-b69966796-j6sqr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:19 +0000 UTC Normal Pod simple-prod-query-b69966796-ldqdw.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:19 +0000 UTC Normal Pod simple-prod-query-b69966796-ldqdw.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:19 +0000 UTC Normal Pod simple-prod-query-b69966796-ldqdw.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:19 +0000 UTC Normal Pod simple-prod-query-b69966796-ldqdw.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:19 +0000 UTC Normal Pod simple-prod-query-b69966796-ldqdw.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:19 +0000 UTC Normal Pod simple-prod-query-b69966796-ldqdw.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:19 +0000 UTC Normal Pod simple-prod-query-b69966796-ldqdw.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:19 +0000 UTC Normal Pod simple-prod-query-b69966796-ldqdw.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:23 +0000 UTC Normal Pod check-span-jsczp Binding Scheduled Successfully assigned kuttl-test-teaching-tick/check-span-jsczp to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:23 +0000 UTC Normal Pod check-span-jsczp AddedInterface Add eth0 [10.128.2.24/23] from ovn-kubernetes multus logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:23 +0000 UTC Normal Pod check-span-jsczp.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:23 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-jsczp job-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:23 +0000 UTC Normal Pod report-span-xghpm Binding Scheduled Successfully assigned kuttl-test-teaching-tick/report-span-xghpm to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:23 +0000 UTC Normal Pod report-span-xghpm AddedInterface Add eth0 [10.131.0.43/23] from ovn-kubernetes multus logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:23 +0000 UTC Normal Pod report-span-xghpm.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:23 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-xghpm job-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:24 +0000 UTC Normal Pod report-span-xghpm.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 338ms (338ms including waiting) kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:24 +0000 UTC Normal Pod report-span-xghpm.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:24 +0000 UTC Normal Pod report-span-xghpm.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:25 +0000 UTC Normal Pod check-span-jsczp.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.7s (1.7s including waiting) kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:25 +0000 UTC Normal Pod check-span-jsczp.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:25 +0000 UTC Normal Pod check-span-jsczp.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:36 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:38 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestteachingticksimpleprod-2-55c698448 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestteachingticksimpleprod-2-55c698nt66j replicaset-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingticksimpleprod-2-55c698nt66j Binding Scheduled Successfully assigned kuttl-test-teaching-tick/elasticsearch-cdm-kuttltestteachingticksimpleprod-2-55c698nt66j to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:38 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestteachingticksimpleprod-2 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestteachingticksimpleprod-2-55c698448 to 1 deployment-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:38 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled down replica set simple-prod-collector-5cd647b47b to 0 from 2 deployment-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:38 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-b69966796 to 0 from 2 deployment-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingticksimpleprod-2-55c698nt66j AddedInterface Add eth0 [10.131.0.44/23] from ovn-kubernetes multus logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingticksimpleprod-2-55c698nt66j.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:39 +0000 UTC Normal Pod simple-prod-collector-5cd647b47b-5hzc7.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:39 +0000 UTC Normal Pod simple-prod-collector-5cd647b47b-r6chb.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:39 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5cd647b47b SuccessfulDelete Deleted pod: simple-prod-collector-5cd647b47b-r6chb replicaset-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:39 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5cd647b47b SuccessfulDelete Deleted pod: simple-prod-collector-5cd647b47b-5hzc7 replicaset-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:39 +0000 UTC Normal Pod simple-prod-query-b69966796-j6sqr.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:39 +0000 UTC Normal Pod simple-prod-query-b69966796-j6sqr.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:39 +0000 UTC Normal Pod simple-prod-query-b69966796-j6sqr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:39 +0000 UTC Normal Pod simple-prod-query-b69966796-ldqdw.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:39 +0000 UTC Normal Pod simple-prod-query-b69966796-ldqdw.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:39 +0000 UTC Normal Pod simple-prod-query-b69966796-ldqdw.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:39 +0000 UTC Normal ReplicaSet.apps simple-prod-query-b69966796 SuccessfulDelete Deleted pod: simple-prod-query-b69966796-ldqdw replicaset-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:39 +0000 UTC Normal ReplicaSet.apps simple-prod-query-b69966796 SuccessfulDelete Deleted pod: simple-prod-query-b69966796-j6sqr replicaset-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingticksimpleprod-2-55c698nt66j.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingticksimpleprod-2-55c698nt66j.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingticksimpleprod-2-55c698nt66j.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingticksimpleprod-2-55c698nt66j.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingticksimpleprod-2-55c698nt66j.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:40 +0000 UTC Normal ReplicaSet.apps simple-prod-query-9c59fbc89 SuccessfulCreate Created pod: simple-prod-query-9c59fbc89-pczpk replicaset-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:40 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-9c59fbc89 to 2 deployment-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:41 +0000 UTC Normal Pod simple-prod-collector-6c4b5c47f9-jwlhp Binding Scheduled Successfully assigned kuttl-test-teaching-tick/simple-prod-collector-6c4b5c47f9-jwlhp to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:41 +0000 UTC Normal Pod simple-prod-collector-6c4b5c47f9-l8926 Binding Scheduled Successfully assigned kuttl-test-teaching-tick/simple-prod-collector-6c4b5c47f9-l8926 to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:41 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-6c4b5c47f9 SuccessfulCreate Created pod: simple-prod-collector-6c4b5c47f9-l8926 replicaset-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:41 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-6c4b5c47f9 SuccessfulCreate Created pod: simple-prod-collector-6c4b5c47f9-jwlhp replicaset-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:41 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-6c4b5c47f9 to 2 deployment-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:41 +0000 UTC Normal Pod simple-prod-query-9c59fbc89-p9rlk Binding Scheduled Successfully assigned kuttl-test-teaching-tick/simple-prod-query-9c59fbc89-p9rlk to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:41 +0000 UTC Normal Pod simple-prod-query-9c59fbc89-pczpk Binding Scheduled Successfully assigned kuttl-test-teaching-tick/simple-prod-query-9c59fbc89-pczpk to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:41 +0000 UTC Normal Pod simple-prod-query-9c59fbc89-pczpk AddedInterface Add eth0 [10.128.2.25/23] from ovn-kubernetes multus logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:41 +0000 UTC Normal Pod simple-prod-query-9c59fbc89-pczpk.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:41 +0000 UTC Normal ReplicaSet.apps simple-prod-query-9c59fbc89 SuccessfulCreate Created pod: simple-prod-query-9c59fbc89-p9rlk replicaset-controller logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:42 +0000 UTC Normal Pod simple-prod-collector-6c4b5c47f9-jwlhp AddedInterface Add eth0 [10.131.0.46/23] from ovn-kubernetes multus logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:42 +0000 UTC Normal Pod simple-prod-collector-6c4b5c47f9-jwlhp.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:42 +0000 UTC Normal Pod simple-prod-collector-6c4b5c47f9-l8926 AddedInterface Add eth0 [10.128.2.26/23] from ovn-kubernetes multus logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:42 +0000 UTC Normal Pod simple-prod-collector-6c4b5c47f9-l8926.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:42 +0000 UTC Normal Pod simple-prod-collector-6c4b5c47f9-l8926.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:42 +0000 UTC Normal Pod simple-prod-collector-6c4b5c47f9-l8926.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:42 +0000 UTC Normal Pod simple-prod-query-9c59fbc89-p9rlk AddedInterface Add eth0 [10.131.0.45/23] from ovn-kubernetes multus logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:42 +0000 UTC Normal Pod simple-prod-query-9c59fbc89-p9rlk.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:42 +0000 UTC Normal Pod simple-prod-query-9c59fbc89-p9rlk.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:42 +0000 UTC Normal Pod simple-prod-query-9c59fbc89-pczpk.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:42 +0000 UTC Normal Pod simple-prod-query-9c59fbc89-pczpk.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:42 +0000 UTC Normal Pod simple-prod-query-9c59fbc89-pczpk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:42 +0000 UTC Normal Pod simple-prod-query-9c59fbc89-pczpk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:42 +0000 UTC Normal Pod simple-prod-query-9c59fbc89-pczpk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:42 +0000 UTC Normal Pod simple-prod-query-9c59fbc89-pczpk.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:42 +0000 UTC Normal Pod simple-prod-query-9c59fbc89-pczpk.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 15:37:42 | es-increasing-replicas | 2024-06-17 15:37:42 +0000 UTC Normal Pod simple-prod-query-9c59fbc89-pczpk.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 15:37:42 | es-increasing-replicas | Deleting namespace: kuttl-test-teaching-tick === CONT kuttl/harness/es-index-cleaner-autoprov logger.go:42: 15:38:18 | es-index-cleaner-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 15:38:18 | es-index-cleaner-autoprov | Creating namespace: kuttl-test-optimal-emu logger.go:42: 15:38:18 | es-index-cleaner-autoprov/1-install | starting test step 1-install logger.go:42: 15:38:18 | es-index-cleaner-autoprov/1-install | Jaeger:kuttl-test-optimal-emu/test-es-index-cleaner-with-prefix created logger.go:42: 15:38:54 | es-index-cleaner-autoprov/1-install | test step completed 1-install logger.go:42: 15:38:54 | es-index-cleaner-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 15:38:54 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE test-es-index-cleaner-with-prefix /dev/null] logger.go:42: 15:38:56 | es-index-cleaner-autoprov/2-report-spans | Warning: resource jaegers/test-es-index-cleaner-with-prefix is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 15:39:03 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c DAYS=5 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 15:39:03 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 15:39:04 | es-index-cleaner-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 15:49:04 | es-index-cleaner-autoprov/2-report-spans | test step failed 2-report-spans case.go:364: failed in step 2-report-spans case.go:366: --- Job:kuttl-test-optimal-emu/00-report-span +++ Job:kuttl-test-optimal-emu/00-report-span @@ -1,8 +1,162 @@ apiVersion: batch/v1 kind: Job metadata: + annotations: + kubectl.kubernetes.io/last-applied-configuration: | + {"apiVersion":"batch/v1","kind":"Job","metadata":{"annotations":{},"name":"00-report-span","namespace":"kuttl-test-optimal-emu"},"spec":{"backoffLimit":15,"template":{"spec":{"containers":[{"command":["./reporter","--days","5","--verbose"],"env":[{"name":"JAEGER_SERVICE_NAME","value":"smoke-test-service"},{"name":"OPERATION_NAME","value":"smoke-test-operation"},{"name":"JAEGER_ENDPOINT","value":"http://test-es-index-cleaner-with-prefix-collector-headless:14268/api/traces"},{"name":"JAEGER_QUERY","value":"https://test-es-index-cleaner-with-prefix-query/api/traces"},{"name":"SECRET_PATH","value":"/var/run/secrets/api-token/token"}],"image":"quay.io/rhn_support_ikanse/jaeger-asserts:latest","name":"asserts-container","volumeMounts":[{"mountPath":"/var/run/secrets/api-token","name":"token-api-volume"}]}],"restartPolicy":"OnFailure","volumes":[{"name":"token-api-volume","secret":{"secretName":"e2e-test"}}]}}}} + labels: + batch.kubernetes.io/controller-uid: d2e49938-a816-47c1-9c11-c26f42ea8503 + batch.kubernetes.io/job-name: 00-report-span + controller-uid: d2e49938-a816-47c1-9c11-c26f42ea8503 + job-name: 00-report-span + managedFields: + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:kubectl.kubernetes.io/last-applied-configuration: {} + f:spec: + f:backoffLimit: {} + f:completionMode: {} + f:completions: {} + f:manualSelector: {} + f:parallelism: {} + f:podReplacementPolicy: {} + f:suspend: {} + f:template: + f:spec: + f:containers: + k:{"name":"asserts-container"}: + .: {} + f:command: {} + f:env: + .: {} + k:{"name":"JAEGER_ENDPOINT"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"JAEGER_QUERY"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"JAEGER_SERVICE_NAME"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"OPERATION_NAME"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SECRET_PATH"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:name: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/var/run/secrets/api-token"}: + .: {} + f:mountPath: {} + f:name: {} + f:dnsPolicy: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"token-api-volume"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + manager: kubectl-client-side-apply + operation: Update + time: "2024-06-17T15:39:04Z" + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:status: + f:active: {} + f:ready: {} + f:startTime: {} + f:terminating: {} + f:uncountedTerminatedPods: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-06-17T15:39:08Z" name: 00-report-span namespace: kuttl-test-optimal-emu +spec: + backoffLimit: 15 + completionMode: NonIndexed + completions: 1 + manualSelector: false + parallelism: 1 + podReplacementPolicy: TerminatingOrFailed + selector: + matchLabels: + batch.kubernetes.io/controller-uid: d2e49938-a816-47c1-9c11-c26f42ea8503 + suspend: false + template: + metadata: + creationTimestamp: null + labels: + batch.kubernetes.io/controller-uid: d2e49938-a816-47c1-9c11-c26f42ea8503 + batch.kubernetes.io/job-name: 00-report-span + controller-uid: d2e49938-a816-47c1-9c11-c26f42ea8503 + job-name: 00-report-span + spec: + containers: + - command: + - ./reporter + - --days + - "5" + - --verbose + env: + - name: JAEGER_SERVICE_NAME + value: smoke-test-service + - name: OPERATION_NAME + value: smoke-test-operation + - name: JAEGER_ENDPOINT + value: http://test-es-index-cleaner-with-prefix-collector-headless:14268/api/traces + - name: JAEGER_QUERY + value: https://test-es-index-cleaner-with-prefix-query/api/traces + - name: SECRET_PATH + value: /var/run/secrets/api-token/token + image: quay.io/rhn_support_ikanse/jaeger-asserts:latest + imagePullPolicy: Always + name: asserts-container + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /var/run/secrets/api-token + name: token-api-volume + dnsPolicy: ClusterFirst + restartPolicy: OnFailure + schedulerName: default-scheduler + securityContext: {} + terminationGracePeriodSeconds: 30 + volumes: + - name: token-api-volume + secret: + defaultMode: 420 + secretName: e2e-test status: - succeeded: 1 + active: 1 + ready: 1 + startTime: "2024-06-17T15:39:04Z" + terminating: 0 + uncountedTerminatedPods: {} case.go:366: resource Job:kuttl-test-optimal-emu/00-report-span: .status.succeeded: key is missing from map logger.go:42: 15:49:04 | es-index-cleaner-autoprov | es-index-cleaner-autoprov events from ns kuttl-test-optimal-emu: logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:24 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestoptimalemutestesindexcleane-1-7549cddc94 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestoptimalemutestesindexcleane-1-7xzrmd replicaset-controller logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:24 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestoptimalemutestesindexcleane-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestoptimalemutestesindexcleane-1-7549cddc94 to 1 deployment-controller logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestoptimalemutestesindexcleane-1-7xzrmd Binding Scheduled Successfully assigned kuttl-test-optimal-emu/elasticsearch-cdm-kuttltestoptimalemutestesindexcleane-1-7xzrmd to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestoptimalemutestesindexcleane-1-7xzrmd AddedInterface Add eth0 [10.129.2.23/23] from ovn-kubernetes multus logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestoptimalemutestesindexcleane-1-7xzrmd.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestoptimalemutestesindexcleane-1-7xzrmd.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestoptimalemutestesindexcleane-1-7xzrmd.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestoptimalemutestesindexcleane-1-7xzrmd.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestoptimalemutestesindexcleane-1-7xzrmd.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestoptimalemutestesindexcleane-1-7xzrmd.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:35 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestoptimalemutestesindexcleane-1-7xzrmd.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:40 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestoptimalemutestesindexcleane-1-7xzrmd.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:51 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-5cb946d5f-4flmg Binding Scheduled Successfully assigned kuttl-test-optimal-emu/test-es-index-cleaner-with-prefix-collector-5cb946d5f-4flmg to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:51 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-collector-5cb946d5f SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-collector-5cb946d5f-4flmg replicaset-controller logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:51 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-collector ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-collector-5cb946d5f to 1 deployment-controller logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:51 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ff4b46b5-lqq64 Binding Scheduled Successfully assigned kuttl-test-optimal-emu/test-es-index-cleaner-with-prefix-query-77ff4b46b5-lqq64 to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:51 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-77ff4b46b5 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-77ff4b46b5-lqq64 replicaset-controller logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:51 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-77ff4b46b5 to 1 deployment-controller logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-5cb946d5f-4flmg AddedInterface Add eth0 [10.131.0.47/23] from ovn-kubernetes multus logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-5cb946d5f-4flmg.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1" already present on machine kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-5cb946d5f-4flmg.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-5cb946d5f-4flmg.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ff4b46b5-lqq64 AddedInterface Add eth0 [10.128.2.27/23] from ovn-kubernetes multus logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ff4b46b5-lqq64.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ff4b46b5-lqq64.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ff4b46b5-lqq64.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ff4b46b5-lqq64.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ff4b46b5-lqq64.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ff4b46b5-lqq64.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ff4b46b5-lqq64.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ff4b46b5-lqq64.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:38:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ff4b46b5-lqq64.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:39:04 +0000 UTC Normal Pod 00-report-span-445xl Binding Scheduled Successfully assigned kuttl-test-optimal-emu/00-report-span-445xl to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:39:04 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-445xl job-controller logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:39:05 +0000 UTC Normal Pod 00-report-span-445xl AddedInterface Add eth0 [10.131.0.48/23] from ovn-kubernetes multus logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:39:05 +0000 UTC Normal Pod 00-report-span-445xl.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:39:06 +0000 UTC Normal Pod 00-report-span-445xl.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.194s (1.194s including waiting) kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:39:06 +0000 UTC Normal Pod 00-report-span-445xl.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:39:06 +0000 UTC Normal Pod 00-report-span-445xl.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:39:09 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:39:09 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-5cb946d5f-4flmg horizontal-pod-autoscaler logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:39:09 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:39:24 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-5cb946d5f-4flmg horizontal-pod-autoscaler logger.go:42: 15:49:04 | es-index-cleaner-autoprov | 2024-06-17 15:39:24 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-5cb946d5f-4flmg horizontal-pod-autoscaler logger.go:42: 15:49:04 | es-index-cleaner-autoprov | Deleting namespace: kuttl-test-optimal-emu === CONT kuttl/harness/es-from-aio-to-production logger.go:42: 15:49:16 | es-from-aio-to-production | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 15:49:16 | es-from-aio-to-production | Creating namespace: kuttl-test-trusting-prawn logger.go:42: 15:49:16 | es-from-aio-to-production/0-install | starting test step 0-install logger.go:42: 15:49:16 | es-from-aio-to-production/0-install | Jaeger:kuttl-test-trusting-prawn/my-jaeger created logger.go:42: 15:49:27 | es-from-aio-to-production/0-install | test step completed 0-install logger.go:42: 15:49:27 | es-from-aio-to-production/1-smoke-test | starting test step 1-smoke-test logger.go:42: 15:49:27 | es-from-aio-to-production/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 15:49:28 | es-from-aio-to-production/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 15:49:35 | es-from-aio-to-production/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 15:49:36 | es-from-aio-to-production/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 15:49:36 | es-from-aio-to-production/1-smoke-test | job.batch/report-span created logger.go:42: 15:49:36 | es-from-aio-to-production/1-smoke-test | job.batch/check-span created logger.go:42: 15:49:49 | es-from-aio-to-production/1-smoke-test | test step completed 1-smoke-test logger.go:42: 15:49:49 | es-from-aio-to-production/3-install | starting test step 3-install logger.go:42: 15:49:49 | es-from-aio-to-production/3-install | Jaeger:kuttl-test-trusting-prawn/my-jaeger updated logger.go:42: 15:50:23 | es-from-aio-to-production/3-install | test step completed 3-install logger.go:42: 15:50:23 | es-from-aio-to-production/4-smoke-test | starting test step 4-smoke-test logger.go:42: 15:50:23 | es-from-aio-to-production/4-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 15:50:31 | es-from-aio-to-production/4-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 15:50:31 | es-from-aio-to-production/4-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 15:50:32 | es-from-aio-to-production/4-smoke-test | job.batch/report-span unchanged logger.go:42: 15:50:32 | es-from-aio-to-production/4-smoke-test | job.batch/check-span unchanged logger.go:42: 15:50:32 | es-from-aio-to-production/4-smoke-test | test step completed 4-smoke-test logger.go:42: 15:50:32 | es-from-aio-to-production | es-from-aio-to-production events from ns kuttl-test-trusting-prawn: logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:20 +0000 UTC Normal Pod my-jaeger-7cc96cd7b5-z2p8n Binding Scheduled Successfully assigned kuttl-test-trusting-prawn/my-jaeger-7cc96cd7b5-z2p8n to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:20 +0000 UTC Normal ReplicaSet.apps my-jaeger-7cc96cd7b5 SuccessfulCreate Created pod: my-jaeger-7cc96cd7b5-z2p8n replicaset-controller logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:20 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-7cc96cd7b5 to 1 deployment-controller logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:21 +0000 UTC Normal Pod my-jaeger-7cc96cd7b5-z2p8n AddedInterface Add eth0 [10.129.2.24/23] from ovn-kubernetes multus logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:21 +0000 UTC Normal Pod my-jaeger-7cc96cd7b5-z2p8n.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:25 +0000 UTC Normal Pod my-jaeger-7cc96cd7b5-z2p8n.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" in 4.374s (4.374s including waiting) kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:25 +0000 UTC Normal Pod my-jaeger-7cc96cd7b5-z2p8n.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:25 +0000 UTC Normal Pod my-jaeger-7cc96cd7b5-z2p8n.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:25 +0000 UTC Normal Pod my-jaeger-7cc96cd7b5-z2p8n.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:25 +0000 UTC Normal Pod my-jaeger-7cc96cd7b5-z2p8n.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:25 +0000 UTC Normal Pod my-jaeger-7cc96cd7b5-z2p8n.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:33 +0000 UTC Normal ReplicaSet.apps my-jaeger-5b5597f5dc SuccessfulCreate Created pod: my-jaeger-5b5597f5dc-snvd6 replicaset-controller logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:33 +0000 UTC Normal Pod my-jaeger-7cc96cd7b5-z2p8n.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:33 +0000 UTC Normal Pod my-jaeger-7cc96cd7b5-z2p8n.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:33 +0000 UTC Normal ReplicaSet.apps my-jaeger-7cc96cd7b5 SuccessfulDelete Deleted pod: my-jaeger-7cc96cd7b5-z2p8n replicaset-controller logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:33 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-7cc96cd7b5 to 0 from 1 deployment-controller logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:33 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5b5597f5dc to 1 deployment-controller logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:34 +0000 UTC Normal Pod my-jaeger-5b5597f5dc-snvd6 Binding Scheduled Successfully assigned kuttl-test-trusting-prawn/my-jaeger-5b5597f5dc-snvd6 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:34 +0000 UTC Normal Pod my-jaeger-5b5597f5dc-snvd6 AddedInterface Add eth0 [10.129.2.25/23] from ovn-kubernetes multus logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:34 +0000 UTC Normal Pod my-jaeger-5b5597f5dc-snvd6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:34 +0000 UTC Normal Pod my-jaeger-5b5597f5dc-snvd6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:34 +0000 UTC Normal Pod my-jaeger-5b5597f5dc-snvd6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:34 +0000 UTC Normal Pod my-jaeger-5b5597f5dc-snvd6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:34 +0000 UTC Normal Pod my-jaeger-5b5597f5dc-snvd6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:34 +0000 UTC Normal Pod my-jaeger-5b5597f5dc-snvd6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:36 +0000 UTC Normal Pod check-span-wr6c2 Binding Scheduled Successfully assigned kuttl-test-trusting-prawn/check-span-wr6c2 to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:36 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-wr6c2 job-controller logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:36 +0000 UTC Normal Pod report-span-8gvjl Binding Scheduled Successfully assigned kuttl-test-trusting-prawn/report-span-8gvjl to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:36 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-8gvjl job-controller logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:37 +0000 UTC Normal Pod check-span-wr6c2 AddedInterface Add eth0 [10.128.2.28/23] from ovn-kubernetes multus logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:37 +0000 UTC Normal Pod check-span-wr6c2.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:37 +0000 UTC Normal Pod check-span-wr6c2.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 348ms (348ms including waiting) kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:37 +0000 UTC Normal Pod check-span-wr6c2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:37 +0000 UTC Normal Pod check-span-wr6c2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:37 +0000 UTC Normal Pod report-span-8gvjl AddedInterface Add eth0 [10.131.0.50/23] from ovn-kubernetes multus logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:37 +0000 UTC Normal Pod report-span-8gvjl.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:37 +0000 UTC Normal Pod report-span-8gvjl.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 378ms (378ms including waiting) kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:37 +0000 UTC Normal Pod report-span-8gvjl.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:37 +0000 UTC Normal Pod report-span-8gvjl.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:48 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:53 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltesttrustingprawnmyjaeger-1-5db6b585c8 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltesttrustingprawnmyjaeger-1-5db6b58k9jnq replicaset-controller logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttrustingprawnmyjaeger-1-5db6b58k9jnq Binding Scheduled Successfully assigned kuttl-test-trusting-prawn/elasticsearch-cdm-kuttltesttrustingprawnmyjaeger-1-5db6b58k9jnq to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:53 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltesttrustingprawnmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltesttrustingprawnmyjaeger-1-5db6b585c8 to 1 deployment-controller logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttrustingprawnmyjaeger-1-5db6b58k9jnq AddedInterface Add eth0 [10.131.0.51/23] from ovn-kubernetes multus logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttrustingprawnmyjaeger-1-5db6b58k9jnq.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttrustingprawnmyjaeger-1-5db6b58k9jnq.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttrustingprawnmyjaeger-1-5db6b58k9jnq.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttrustingprawnmyjaeger-1-5db6b58k9jnq.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttrustingprawnmyjaeger-1-5db6b58k9jnq.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:49:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttrustingprawnmyjaeger-1-5db6b58k9jnq.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:04 +0000 UTC Warning Pod elasticsearch-cdm-kuttltesttrustingprawnmyjaeger-1-5db6b58k9jnq.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:09 +0000 UTC Warning Pod elasticsearch-cdm-kuttltesttrustingprawnmyjaeger-1-5db6b58k9jnq.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:11 +0000 UTC Normal Job.batch report-span Completed Job completed job-controller logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:20 +0000 UTC Normal Pod my-jaeger-5b5597f5dc-snvd6.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:20 +0000 UTC Normal Pod my-jaeger-5b5597f5dc-snvd6.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:20 +0000 UTC Normal Pod my-jaeger-collector-7b87b67c8d-xhcq6 Binding Scheduled Successfully assigned kuttl-test-trusting-prawn/my-jaeger-collector-7b87b67c8d-xhcq6 to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:20 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7b87b67c8d SuccessfulCreate Created pod: my-jaeger-collector-7b87b67c8d-xhcq6 replicaset-controller logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:20 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7b87b67c8d to 1 deployment-controller logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:20 +0000 UTC Normal Pod my-jaeger-query-6f9469f484-tbqsx Binding Scheduled Successfully assigned kuttl-test-trusting-prawn/my-jaeger-query-6f9469f484-tbqsx to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:20 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-6f9469f484 SuccessfulCreate Created pod: my-jaeger-query-6f9469f484-tbqsx replicaset-controller logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:20 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-6f9469f484 to 1 deployment-controller logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:21 +0000 UTC Normal Pod my-jaeger-collector-7b87b67c8d-xhcq6 AddedInterface Add eth0 [10.128.2.29/23] from ovn-kubernetes multus logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:21 +0000 UTC Normal Pod my-jaeger-collector-7b87b67c8d-xhcq6.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1" already present on machine kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:21 +0000 UTC Normal Pod my-jaeger-collector-7b87b67c8d-xhcq6.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:21 +0000 UTC Normal Pod my-jaeger-collector-7b87b67c8d-xhcq6.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:21 +0000 UTC Normal Pod my-jaeger-query-6f9469f484-tbqsx AddedInterface Add eth0 [10.128.2.30/23] from ovn-kubernetes multus logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:21 +0000 UTC Normal Pod my-jaeger-query-6f9469f484-tbqsx.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:21 +0000 UTC Normal Pod my-jaeger-query-6f9469f484-tbqsx.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:21 +0000 UTC Normal Pod my-jaeger-query-6f9469f484-tbqsx.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:21 +0000 UTC Normal Pod my-jaeger-query-6f9469f484-tbqsx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:21 +0000 UTC Normal Pod my-jaeger-query-6f9469f484-tbqsx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:21 +0000 UTC Normal Pod my-jaeger-query-6f9469f484-tbqsx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:21 +0000 UTC Normal Pod my-jaeger-query-6f9469f484-tbqsx.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:21 +0000 UTC Normal Pod my-jaeger-query-6f9469f484-tbqsx.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | 2024-06-17 15:50:21 +0000 UTC Normal Pod my-jaeger-query-6f9469f484-tbqsx.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 15:50:32 | es-from-aio-to-production | Deleting namespace: kuttl-test-trusting-prawn === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (1218.85s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.21s) --- PASS: kuttl/harness/es-multiinstance (128.12s) --- PASS: kuttl/harness/es-simple-prod (6.44s) --- PASS: kuttl/harness/es-rollover-autoprov (227.18s) --- PASS: kuttl/harness/es-increasing-replicas (110.01s) --- FAIL: kuttl/harness/es-index-cleaner-autoprov (658.28s) --- PASS: kuttl/harness/es-from-aio-to-production (82.41s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name elasticsearch --report --output /logs/artifacts/elasticsearch.xml ./artifacts/kuttl-report.xml time="2024-06-17T15:50:40Z" level=debug msg="Setting a new name for the test suites" time="2024-06-17T15:50:40Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-06-17T15:50:40Z" level=debug msg="normalizing test case names" time="2024-06-17T15:50:40Z" level=debug msg="elasticsearch/artifacts -> elasticsearch_artifacts" time="2024-06-17T15:50:40Z" level=debug msg="elasticsearch/es-multiinstance -> elasticsearch_es_multiinstance" time="2024-06-17T15:50:40Z" level=debug msg="elasticsearch/es-simple-prod -> elasticsearch_es_simple_prod" time="2024-06-17T15:50:40Z" level=debug msg="elasticsearch/es-rollover-autoprov -> elasticsearch_es_rollover_autoprov" time="2024-06-17T15:50:40Z" level=debug msg="elasticsearch/es-increasing-replicas -> elasticsearch_es_increasing_replicas" time="2024-06-17T15:50:40Z" level=debug msg="elasticsearch/es-index-cleaner-autoprov -> elasticsearch_es_index_cleaner_autoprov" time="2024-06-17T15:50:40Z" level=debug msg="elasticsearch/es-from-aio-to-production -> elasticsearch_es_from_aio_to_production" +-----------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------+--------+ | elasticsearch_artifacts | passed | | elasticsearch_es_multiinstance | passed | | elasticsearch_es_simple_prod | passed | | elasticsearch_es_rollover_autoprov | passed | | elasticsearch_es_increasing_replicas | passed | | elasticsearch_es_index_cleaner_autoprov | failed | | elasticsearch_es_from_aio_to_production | passed | +-----------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh examples false true + '[' 3 -ne 3 ']' + test_suite_name=examples + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/examples.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-examples make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ VERTX_IMG=jaegertracing/vertx-create-span:operator-e2e-tests \ ./tests/e2e/examples/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-14-130320 True False 32m Cluster version is 4.16.0-0.nightly-2024-06-14-130320' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-14-130320 True False 32m Cluster version is 4.16.0-0.nightly-2024-06-14-130320' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/examples/render.sh ++ export SUITE_DIR=./tests/e2e/examples ++ SUITE_DIR=./tests/e2e/examples ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/examples ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test examples-agent-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-agent-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-with-priority-class\e[0m' Rendering files for test examples-agent-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + mkdir -p examples-agent-with-priority-class + cd examples-agent-with-priority-class + example_name=agent-with-priority-class + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + '[' true '!=' true ']' + render_install_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + install_file=./02-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/agent-with-priority-class.yaml -o ./02-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./02-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./02-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./02-install.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + JAEGER_NAME=agent-as-daemonset + local jaeger_strategy ++ get_jaeger_strategy ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./02-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./02-install.yaml ++ strategy=DaemonSet ++ '[' DaemonSet = null ']' ++ echo DaemonSet ++ return 0 + jaeger_strategy=DaemonSet + '[' DaemonSet = DaemonSet ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./02-assert.yaml + render_smoke_test_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + jaeger_name=agent-as-daemonset + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test agent-as-daemonset true 02 + '[' 3 -ne 3 ']' + jaeger=agent-as-daemonset + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-all-in-one-with-options + '[' 1 -ne 1 ']' + test_name=examples-all-in-one-with-options + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-all-in-one-with-options' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-all-in-one-with-options\e[0m' Rendering files for test examples-all-in-one-with-options + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-with-priority-class + '[' examples-agent-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-all-in-one-with-options + cd examples-all-in-one-with-options + example_name=all-in-one-with-options + render_install_example all-in-one-with-options 00 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/all-in-one-with-options.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + JAEGER_NAME=my-jaeger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.metadata.name="my-jaeger"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i 'del(.spec.allInOne.image)' ./00-install.yaml + render_smoke_test_example all-in-one-with-options 01 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + jaeger_name=my-jaeger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + sed -i s~my-jaeger-query:443~my-jaeger-query:443/jaeger~gi ./01-smoke-test.yaml + start_test examples-business-application-injected-sidecar + '[' 1 -ne 1 ']' + test_name=examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-business-application-injected-sidecar' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-business-application-injected-sidecar\e[0m' Rendering files for test examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-all-in-one-with-options + '[' examples-all-in-one-with-options '!=' _build ']' + cd .. + mkdir -p examples-business-application-injected-sidecar + cd examples-business-application-injected-sidecar + example_name=simplest + cp /tmp/jaeger-tests/examples/business-application-injected-sidecar.yaml ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].image=strenv(VERTX_IMG)' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.path="/"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.port=8080' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.initialDelaySeconds=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.failureThreshold=3' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.periodSeconds=10' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.successThreshold=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.timeoutSeconds=1' ./00-install.yaml + render_install_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example simplest 02 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 02 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-collector-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-collector-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-collector-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-collector-with-priority-class\e[0m' Rendering files for test examples-collector-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-business-application-injected-sidecar + '[' examples-business-application-injected-sidecar '!=' _build ']' + cd .. + mkdir -p examples-collector-with-priority-class + cd examples-collector-with-priority-class + example_name=collector-with-priority-class + render_install_example collector-with-priority-class 00 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/collector-with-priority-class.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + JAEGER_NAME=collector-with-high-priority + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example collector-with-priority-class 01 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + jaeger_name=collector-with-high-priority + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test collector-with-high-priority true 01 + '[' 3 -ne 3 ']' + jaeger=collector-with-high-priority + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + export JAEGER_NAME=collector-with-high-priority + JAEGER_NAME=collector-with-high-priority + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-service-types + '[' 1 -ne 1 ']' + test_name=examples-service-types + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-service-types' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-service-types\e[0m' Rendering files for test examples-service-types + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-collector-with-priority-class + '[' examples-collector-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-service-types + cd examples-service-types + example_name=service-types + render_install_example service-types 00 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/service-types.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + JAEGER_NAME=service-types + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example service-types 01 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/service-types.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/service-types.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/service-types.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/service-types.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + jaeger_name=service-types + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test service-types true 01 + '[' 3 -ne 3 ']' + jaeger=service-types + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + export JAEGER_NAME=service-types + JAEGER_NAME=service-types + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod + '[' 1 -ne 1 ']' + test_name=examples-simple-prod + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod\e[0m' Rendering files for test examples-simple-prod + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-service-types + '[' examples-service-types '!=' _build ']' + cd .. + mkdir -p examples-simple-prod + cd examples-simple-prod + example_name=simple-prod + render_install_example simple-prod 01 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod 02 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod-with-volumes + '[' 1 -ne 1 ']' + test_name=examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod-with-volumes' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod-with-volumes\e[0m' Rendering files for test examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod + '[' examples-simple-prod '!=' _build ']' + cd .. + mkdir -p examples-simple-prod-with-volumes + cd examples-simple-prod-with-volumes + example_name=simple-prod-with-volumes + render_install_example simple-prod-with-volumes 01 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod-with-volumes 02 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + /tmp/jaeger-tests/bin/gomplate -f ./03-check-volume.yaml.template -o 03-check-volume.yaml + start_test examples-simplest + '[' 1 -ne 1 ']' + test_name=examples-simplest + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simplest' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simplest\e[0m' Rendering files for test examples-simplest + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod-with-volumes + '[' examples-simple-prod-with-volumes '!=' _build ']' + cd .. + mkdir -p examples-simplest + cd examples-simplest + example_name=simplest + render_install_example simplest 00 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 01 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger + '[' 1 -ne 1 ']' + test_name=examples-with-badger + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger\e[0m' Rendering files for test examples-with-badger + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simplest + '[' examples-simplest '!=' _build ']' + cd .. + mkdir -p examples-with-badger + cd examples-with-badger + example_name=with-badger + render_install_example with-badger 00 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + JAEGER_NAME=with-badger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger 01 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + jaeger_name=with-badger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + export JAEGER_NAME=with-badger + JAEGER_NAME=with-badger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger-and-volume + '[' 1 -ne 1 ']' + test_name=examples-with-badger-and-volume + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger-and-volume' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger-and-volume\e[0m' Rendering files for test examples-with-badger-and-volume + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger + '[' examples-with-badger '!=' _build ']' + cd .. + mkdir -p examples-with-badger-and-volume + cd examples-with-badger-and-volume + example_name=with-badger-and-volume + render_install_example with-badger-and-volume 00 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger-and-volume.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + JAEGER_NAME=with-badger-and-volume + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger-and-volume 01 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + jaeger_name=with-badger-and-volume + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger-and-volume true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger-and-volume + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + export JAEGER_NAME=with-badger-and-volume + JAEGER_NAME=with-badger-and-volume + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-cassandra + '[' 1 -ne 1 ']' + test_name=examples-with-cassandra + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-cassandra' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-cassandra\e[0m' Rendering files for test examples-with-cassandra + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger-and-volume + '[' examples-with-badger-and-volume '!=' _build ']' + cd .. + mkdir -p examples-with-cassandra + cd examples-with-cassandra + example_name=with-cassandra + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-cassandra 01 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-cassandra.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + JAEGER_NAME=with-cassandra + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-cassandra 02 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-cassandra.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-cassandra.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + jaeger_name=with-cassandra + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-cassandra true 02 + '[' 3 -ne 3 ']' + jaeger=with-cassandra + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + export JAEGER_NAME=with-cassandra + JAEGER_NAME=with-cassandra + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-sampling + '[' 1 -ne 1 ']' + test_name=examples-with-sampling + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-sampling' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-sampling\e[0m' Rendering files for test examples-with-sampling + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-cassandra + '[' examples-with-cassandra '!=' _build ']' + cd .. + mkdir -p examples-with-sampling + cd examples-with-sampling + export example_name=with-sampling + example_name=with-sampling + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-sampling 01 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-sampling.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + JAEGER_NAME=with-sampling + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-sampling 02 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-sampling.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-sampling.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + jaeger_name=with-sampling + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-sampling true 02 + '[' 3 -ne 3 ']' + jaeger=with-sampling + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + export JAEGER_NAME=with-sampling + JAEGER_NAME=with-sampling + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-agent-as-daemonset + '[' 1 -ne 1 ']' + test_name=examples-agent-as-daemonset + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-as-daemonset' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-as-daemonset\e[0m' Rendering files for test examples-agent-as-daemonset + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-sampling + '[' examples-with-sampling '!=' _build ']' + cd .. + mkdir -p examples-agent-as-daemonset + cd examples-agent-as-daemonset + '[' true = true ']' + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/agent-as-daemonset.yaml -o 02-install.yaml + '[' true = true ']' + start_test examples-openshift-with-htpasswd + '[' 1 -ne 1 ']' + test_name=examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-openshift-with-htpasswd' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-openshift-with-htpasswd\e[0m' Rendering files for test examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-as-daemonset + '[' examples-agent-as-daemonset '!=' _build ']' + cd .. + mkdir -p examples-openshift-with-htpasswd + cd examples-openshift-with-htpasswd + export JAEGER_NAME=with-htpasswd + JAEGER_NAME=with-htpasswd + export JAEGER_USERNAME=awesomeuser + JAEGER_USERNAME=awesomeuser + export JAEGER_PASSWORD=awesomepassword + JAEGER_PASSWORD=awesomepassword + export 'JAEGER_USER_PASSWORD_HASH=awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' + JAEGER_USER_PASSWORD_HASH='awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ echo 'awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ base64 + SECRET=YXdlc29tZXVzZXI6e1NIQX11VWRxUFZVeXFOQm1FUlUwUXhqM0tGYVpuanc9Cg== + /tmp/jaeger-tests/bin/gomplate -f ./00-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/with-htpasswd.yaml -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + export 'GET_URL_COMMAND=kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + export 'URL=https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + INSECURE=true + JAEGER_USERNAME= + JAEGER_PASSWORD= + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-unsecured.yaml + JAEGER_USERNAME=wronguser + JAEGER_PASSWORD=wrongpassword + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-check-unauthorized.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./04-check-authorized.yaml + skip_test examples-agent-as-daemonset 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-agent-as-daemonset + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-openshift-with-htpasswd + '[' examples-openshift-with-htpasswd '!=' _build ']' + cd .. + rm -rf examples-agent-as-daemonset + warning 'examples-agent-as-daemonset: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-agent-as-daemonset: This test is flaky in Prow CI\e[0m' WAR: examples-agent-as-daemonset: This test is flaky in Prow CI + skip_test examples-with-badger-and-volume 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-with-badger-and-volume + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-with-badger-and-volume + warning 'examples-with-badger-and-volume: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-with-badger-and-volume: This test is flaky in Prow CI\e[0m' WAR: examples-with-badger-and-volume: This test is flaky in Prow CI + skip_test examples-collector-with-priority-class 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-collector-with-priority-class + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-collector-with-priority-class + warning 'examples-collector-with-priority-class: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-collector-with-priority-class: This test is flaky in Prow CI\e[0m' WAR: examples-collector-with-priority-class: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running examples E2E tests' Running examples E2E tests + cd tests/e2e/examples/_build + set +e + KUBECONFIG=/tmp/kubeconfig-37736524 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 12 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/examples-agent-with-priority-class === PAUSE kuttl/harness/examples-agent-with-priority-class === RUN kuttl/harness/examples-all-in-one-with-options === PAUSE kuttl/harness/examples-all-in-one-with-options === RUN kuttl/harness/examples-business-application-injected-sidecar === PAUSE kuttl/harness/examples-business-application-injected-sidecar === RUN kuttl/harness/examples-openshift-with-htpasswd === PAUSE kuttl/harness/examples-openshift-with-htpasswd === RUN kuttl/harness/examples-service-types === PAUSE kuttl/harness/examples-service-types === RUN kuttl/harness/examples-simple-prod === PAUSE kuttl/harness/examples-simple-prod === RUN kuttl/harness/examples-simple-prod-with-volumes === PAUSE kuttl/harness/examples-simple-prod-with-volumes === RUN kuttl/harness/examples-simplest === PAUSE kuttl/harness/examples-simplest === RUN kuttl/harness/examples-with-badger === PAUSE kuttl/harness/examples-with-badger === RUN kuttl/harness/examples-with-cassandra === PAUSE kuttl/harness/examples-with-cassandra === RUN kuttl/harness/examples-with-sampling === PAUSE kuttl/harness/examples-with-sampling === CONT kuttl/harness/artifacts logger.go:42: 15:51:11 | artifacts | Creating namespace: kuttl-test-polite-rabbit logger.go:42: 15:51:11 | artifacts | artifacts events from ns kuttl-test-polite-rabbit: logger.go:42: 15:51:11 | artifacts | Deleting namespace: kuttl-test-polite-rabbit === CONT kuttl/harness/examples-simple-prod logger.go:42: 15:51:17 | examples-simple-prod | Creating namespace: kuttl-test-pretty-skink logger.go:42: 15:51:17 | examples-simple-prod/1-install | starting test step 1-install logger.go:42: 15:51:17 | examples-simple-prod/1-install | Jaeger:kuttl-test-pretty-skink/simple-prod created logger.go:42: 15:51:54 | examples-simple-prod/1-install | test step completed 1-install logger.go:42: 15:51:54 | examples-simple-prod/2-smoke-test | starting test step 2-smoke-test logger.go:42: 15:51:54 | examples-simple-prod/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 15:51:55 | examples-simple-prod/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 15:52:03 | examples-simple-prod/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 15:52:03 | examples-simple-prod/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 15:52:04 | examples-simple-prod/2-smoke-test | job.batch/report-span created logger.go:42: 15:52:04 | examples-simple-prod/2-smoke-test | job.batch/check-span created logger.go:42: 15:52:16 | examples-simple-prod/2-smoke-test | test step completed 2-smoke-test logger.go:42: 15:52:16 | examples-simple-prod | examples-simple-prod events from ns kuttl-test-pretty-skink: logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:23 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestprettyskinksimpleprod-1-68bb99c5d8 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestprettyskinksimpleprod-1-68bb99cc7ptb replicaset-controller logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprettyskinksimpleprod-1-68bb99cc7ptb Binding Scheduled Successfully assigned kuttl-test-pretty-skink/elasticsearch-cdm-kuttltestprettyskinksimpleprod-1-68bb99cc7ptb to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:23 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestprettyskinksimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestprettyskinksimpleprod-1-68bb99c5d8 to 1 deployment-controller logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprettyskinksimpleprod-1-68bb99cc7ptb AddedInterface Add eth0 [10.129.2.26/23] from ovn-kubernetes multus logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprettyskinksimpleprod-1-68bb99cc7ptb.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprettyskinksimpleprod-1-68bb99cc7ptb.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprettyskinksimpleprod-1-68bb99cc7ptb.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprettyskinksimpleprod-1-68bb99cc7ptb.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprettyskinksimpleprod-1-68bb99cc7ptb.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprettyskinksimpleprod-1-68bb99cc7ptb.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:34 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestprettyskinksimpleprod-1-68bb99cc7ptb.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:39 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestprettyskinksimpleprod-1-68bb99cc7ptb.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:50 +0000 UTC Normal Pod simple-prod-collector-574f4c4fd6-ffqt5 Binding Scheduled Successfully assigned kuttl-test-pretty-skink/simple-prod-collector-574f4c4fd6-ffqt5 to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:50 +0000 UTC Warning Pod simple-prod-collector-574f4c4fd6-ffqt5 FailedMount MountVolume.SetUp failed for volume "simple-prod-collector-tls-config-volume" : secret "simple-prod-collector-headless-tls" not found kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:50 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-574f4c4fd6 SuccessfulCreate Created pod: simple-prod-collector-574f4c4fd6-ffqt5 replicaset-controller logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:50 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-574f4c4fd6 to 1 deployment-controller logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:50 +0000 UTC Normal Pod simple-prod-query-7f84d9968-zsxp8 Binding Scheduled Successfully assigned kuttl-test-pretty-skink/simple-prod-query-7f84d9968-zsxp8 to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:50 +0000 UTC Normal ReplicaSet.apps simple-prod-query-7f84d9968 SuccessfulCreate Created pod: simple-prod-query-7f84d9968-zsxp8 replicaset-controller logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:50 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-7f84d9968 to 1 deployment-controller logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:51 +0000 UTC Normal Pod simple-prod-collector-574f4c4fd6-ffqt5 AddedInterface Add eth0 [10.131.0.52/23] from ovn-kubernetes multus logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:51 +0000 UTC Normal Pod simple-prod-collector-574f4c4fd6-ffqt5.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1" already present on machine kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:51 +0000 UTC Normal Pod simple-prod-collector-574f4c4fd6-ffqt5.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:51 +0000 UTC Normal Pod simple-prod-collector-574f4c4fd6-ffqt5.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:51 +0000 UTC Normal Pod simple-prod-query-7f84d9968-zsxp8 AddedInterface Add eth0 [10.128.2.31/23] from ovn-kubernetes multus logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:51 +0000 UTC Normal Pod simple-prod-query-7f84d9968-zsxp8.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:51 +0000 UTC Normal Pod simple-prod-query-7f84d9968-zsxp8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:51 +0000 UTC Normal Pod simple-prod-query-7f84d9968-zsxp8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:51 +0000 UTC Normal Pod simple-prod-query-7f84d9968-zsxp8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:51 +0000 UTC Normal Pod simple-prod-query-7f84d9968-zsxp8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:51 +0000 UTC Normal Pod simple-prod-query-7f84d9968-zsxp8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:51 +0000 UTC Normal Pod simple-prod-query-7f84d9968-zsxp8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:51 +0000 UTC Normal Pod simple-prod-query-7f84d9968-zsxp8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:51 +0000 UTC Normal Pod simple-prod-query-7f84d9968-zsxp8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:59 +0000 UTC Normal Pod simple-prod-query-7f84d9968-zsxp8.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:59 +0000 UTC Normal Pod simple-prod-query-7f84d9968-zsxp8.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:59 +0000 UTC Normal Pod simple-prod-query-7f84d9968-zsxp8.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:59 +0000 UTC Normal ReplicaSet.apps simple-prod-query-7f84d9968 SuccessfulDelete Deleted pod: simple-prod-query-7f84d9968-zsxp8 replicaset-controller logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:51:59 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-7f84d9968 to 0 from 1 deployment-controller logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:00 +0000 UTC Normal Pod simple-prod-query-5cb6f4d9fc-z277s Binding Scheduled Successfully assigned kuttl-test-pretty-skink/simple-prod-query-5cb6f4d9fc-z277s to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:00 +0000 UTC Normal Pod simple-prod-query-5cb6f4d9fc-z277s AddedInterface Add eth0 [10.128.2.32/23] from ovn-kubernetes multus logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:00 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5cb6f4d9fc SuccessfulCreate Created pod: simple-prod-query-5cb6f4d9fc-z277s replicaset-controller logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:00 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-5cb6f4d9fc to 1 deployment-controller logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:01 +0000 UTC Normal Pod simple-prod-query-5cb6f4d9fc-z277s.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:01 +0000 UTC Normal Pod simple-prod-query-5cb6f4d9fc-z277s.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:01 +0000 UTC Normal Pod simple-prod-query-5cb6f4d9fc-z277s.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:01 +0000 UTC Normal Pod simple-prod-query-5cb6f4d9fc-z277s.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:01 +0000 UTC Normal Pod simple-prod-query-5cb6f4d9fc-z277s.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:01 +0000 UTC Normal Pod simple-prod-query-5cb6f4d9fc-z277s.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:01 +0000 UTC Normal Pod simple-prod-query-5cb6f4d9fc-z277s.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:01 +0000 UTC Normal Pod simple-prod-query-5cb6f4d9fc-z277s.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:01 +0000 UTC Normal Pod simple-prod-query-5cb6f4d9fc-z277s.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:04 +0000 UTC Normal Pod check-span-thkch Binding Scheduled Successfully assigned kuttl-test-pretty-skink/check-span-thkch to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:04 +0000 UTC Normal Pod check-span-thkch AddedInterface Add eth0 [10.131.0.54/23] from ovn-kubernetes multus logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:04 +0000 UTC Normal Pod check-span-thkch.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:04 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-thkch job-controller logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:04 +0000 UTC Normal Pod report-span-w6x97 Binding Scheduled Successfully assigned kuttl-test-pretty-skink/report-span-w6x97 to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:04 +0000 UTC Normal Pod report-span-w6x97 AddedInterface Add eth0 [10.131.0.53/23] from ovn-kubernetes multus logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:04 +0000 UTC Normal Pod report-span-w6x97.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:04 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-w6x97 job-controller logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:05 +0000 UTC Normal Pod check-span-thkch.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 430ms (430ms including waiting) kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:05 +0000 UTC Normal Pod check-span-thkch.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:05 +0000 UTC Normal Pod check-span-thkch.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:05 +0000 UTC Normal Pod report-span-w6x97.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 376ms (376ms including waiting) kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:05 +0000 UTC Normal Pod report-span-w6x97.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:05 +0000 UTC Normal Pod report-span-w6x97.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 15:52:16 | examples-simple-prod | 2024-06-17 15:52:16 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 15:52:16 | examples-simple-prod | Deleting namespace: kuttl-test-pretty-skink === CONT kuttl/harness/examples-with-sampling logger.go:42: 15:52:24 | examples-with-sampling | Creating namespace: kuttl-test-alert-raccoon logger.go:42: 15:52:24 | examples-with-sampling/0-install | starting test step 0-install logger.go:42: 15:52:24 | examples-with-sampling/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 15:52:24 | examples-with-sampling/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 15:52:24 | examples-with-sampling/0-install | >>>> Creating namespace kuttl-test-alert-raccoon logger.go:42: 15:52:24 | examples-with-sampling/0-install | kubectl create namespace kuttl-test-alert-raccoon 2>&1 | grep -v "already exists" || true logger.go:42: 15:52:24 | examples-with-sampling/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-alert-raccoon 2>&1 | grep -v "already exists" || true logger.go:42: 15:52:25 | examples-with-sampling/0-install | service/cassandra created logger.go:42: 15:52:25 | examples-with-sampling/0-install | statefulset.apps/cassandra created logger.go:42: 15:52:25 | examples-with-sampling/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 15:52:36 | examples-with-sampling/0-install | test step completed 0-install logger.go:42: 15:52:36 | examples-with-sampling/1-install | starting test step 1-install logger.go:42: 15:52:36 | examples-with-sampling/1-install | Jaeger:kuttl-test-alert-raccoon/with-sampling created logger.go:42: 15:52:42 | examples-with-sampling/1-install | test step completed 1-install logger.go:42: 15:52:42 | examples-with-sampling/2-smoke-test | starting test step 2-smoke-test logger.go:42: 15:52:42 | examples-with-sampling/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-sampling /dev/null] logger.go:42: 15:52:44 | examples-with-sampling/2-smoke-test | Warning: resource jaegers/with-sampling is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 15:52:50 | examples-with-sampling/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 15:52:51 | examples-with-sampling/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 15:52:51 | examples-with-sampling/2-smoke-test | job.batch/report-span created logger.go:42: 15:52:51 | examples-with-sampling/2-smoke-test | job.batch/check-span created logger.go:42: 15:52:58 | examples-with-sampling/2-smoke-test | test step completed 2-smoke-test logger.go:42: 15:52:58 | examples-with-sampling/3- | starting test step 3- logger.go:42: 15:52:58 | examples-with-sampling/3- | test step completed 3- logger.go:42: 15:52:59 | examples-with-sampling | examples-with-sampling events from ns kuttl-test-alert-raccoon: logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:25 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-alert-raccoon/cassandra-0 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:25 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.129.2.27/23] from ovn-kubernetes multus logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:25 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:25 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:29 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 4.246s (4.246s including waiting) kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:30 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:30 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:30 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-alert-raccoon/cassandra-1 to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:30 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:31 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.131.0.55/23] from ovn-kubernetes multus logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:31 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:35 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 4.324s (4.324s including waiting) kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:35 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:35 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:40 +0000 UTC Normal Pod with-sampling-86f77c9ff9-fn2wz Binding Scheduled Successfully assigned kuttl-test-alert-raccoon/with-sampling-86f77c9ff9-fn2wz to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:40 +0000 UTC Normal ReplicaSet.apps with-sampling-86f77c9ff9 SuccessfulCreate Created pod: with-sampling-86f77c9ff9-fn2wz replicaset-controller logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:40 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-86f77c9ff9 to 1 deployment-controller logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:41 +0000 UTC Normal Pod with-sampling-86f77c9ff9-fn2wz AddedInterface Add eth0 [10.129.2.28/23] from ovn-kubernetes multus logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:41 +0000 UTC Normal Pod with-sampling-86f77c9ff9-fn2wz.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:41 +0000 UTC Normal Pod with-sampling-86f77c9ff9-fn2wz.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:41 +0000 UTC Normal Pod with-sampling-86f77c9ff9-fn2wz.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:41 +0000 UTC Normal Pod with-sampling-86f77c9ff9-fn2wz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:41 +0000 UTC Normal Pod with-sampling-86f77c9ff9-fn2wz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:41 +0000 UTC Normal Pod with-sampling-86f77c9ff9-fn2wz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:45 +0000 UTC Normal Pod with-sampling-86f77c9ff9-fn2wz.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:45 +0000 UTC Normal Pod with-sampling-86f77c9ff9-fn2wz.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:45 +0000 UTC Normal ReplicaSet.apps with-sampling-86f77c9ff9 SuccessfulDelete Deleted pod: with-sampling-86f77c9ff9-fn2wz replicaset-controller logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:45 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled down replica set with-sampling-86f77c9ff9 to 0 from 1 deployment-controller logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:46 +0000 UTC Normal Pod with-sampling-68bcb7789b-qqncx Binding Scheduled Successfully assigned kuttl-test-alert-raccoon/with-sampling-68bcb7789b-qqncx to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:46 +0000 UTC Normal ReplicaSet.apps with-sampling-68bcb7789b SuccessfulCreate Created pod: with-sampling-68bcb7789b-qqncx replicaset-controller logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:46 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-68bcb7789b to 1 deployment-controller logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:47 +0000 UTC Normal Pod with-sampling-68bcb7789b-qqncx AddedInterface Add eth0 [10.129.2.29/23] from ovn-kubernetes multus logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:47 +0000 UTC Normal Pod with-sampling-68bcb7789b-qqncx.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:47 +0000 UTC Normal Pod with-sampling-68bcb7789b-qqncx.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:47 +0000 UTC Normal Pod with-sampling-68bcb7789b-qqncx.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:47 +0000 UTC Normal Pod with-sampling-68bcb7789b-qqncx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:47 +0000 UTC Normal Pod with-sampling-68bcb7789b-qqncx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:47 +0000 UTC Normal Pod with-sampling-68bcb7789b-qqncx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:51 +0000 UTC Normal Pod check-span-jf695 Binding Scheduled Successfully assigned kuttl-test-alert-raccoon/check-span-jf695 to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:51 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-jf695 job-controller logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:51 +0000 UTC Normal Pod report-span-plkr4 Binding Scheduled Successfully assigned kuttl-test-alert-raccoon/report-span-plkr4 to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:51 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-plkr4 job-controller logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:52 +0000 UTC Normal Pod check-span-jf695 AddedInterface Add eth0 [10.131.0.56/23] from ovn-kubernetes multus logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:52 +0000 UTC Normal Pod check-span-jf695.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:52 +0000 UTC Normal Pod report-span-plkr4 AddedInterface Add eth0 [10.128.2.33/23] from ovn-kubernetes multus logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:52 +0000 UTC Normal Pod report-span-plkr4.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:52 +0000 UTC Normal Pod report-span-plkr4.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 408ms (408ms including waiting) kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:52 +0000 UTC Normal Pod report-span-plkr4.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:52 +0000 UTC Normal Pod report-span-plkr4.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:55 +0000 UTC Normal Pod check-span-jf695.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 3.169s (3.169s including waiting) kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:55 +0000 UTC Normal Pod check-span-jf695.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:55 +0000 UTC Normal Pod check-span-jf695.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 15:52:59 | examples-with-sampling | 2024-06-17 15:52:58 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 15:52:59 | examples-with-sampling | Deleting namespace: kuttl-test-alert-raccoon === CONT kuttl/harness/examples-with-cassandra logger.go:42: 15:57:44 | examples-with-cassandra | Creating namespace: kuttl-test-funny-airedale logger.go:42: 15:57:44 | examples-with-cassandra/0-install | starting test step 0-install logger.go:42: 15:57:44 | examples-with-cassandra/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 15:57:44 | examples-with-cassandra/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 15:57:44 | examples-with-cassandra/0-install | >>>> Creating namespace kuttl-test-funny-airedale logger.go:42: 15:57:44 | examples-with-cassandra/0-install | kubectl create namespace kuttl-test-funny-airedale 2>&1 | grep -v "already exists" || true logger.go:42: 15:57:44 | examples-with-cassandra/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-funny-airedale 2>&1 | grep -v "already exists" || true logger.go:42: 15:57:45 | examples-with-cassandra/0-install | service/cassandra created logger.go:42: 15:57:45 | examples-with-cassandra/0-install | statefulset.apps/cassandra created logger.go:42: 15:57:45 | examples-with-cassandra/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 15:58:08 | examples-with-cassandra/0-install | test step completed 0-install logger.go:42: 15:58:08 | examples-with-cassandra/1-install | starting test step 1-install logger.go:42: 15:58:08 | examples-with-cassandra/1-install | Jaeger:kuttl-test-funny-airedale/with-cassandra created logger.go:42: 15:58:26 | examples-with-cassandra/1-install | test step completed 1-install logger.go:42: 15:58:26 | examples-with-cassandra/2-smoke-test | starting test step 2-smoke-test logger.go:42: 15:58:26 | examples-with-cassandra/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-cassandra /dev/null] logger.go:42: 15:58:27 | examples-with-cassandra/2-smoke-test | Warning: resource jaegers/with-cassandra is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 15:58:34 | examples-with-cassandra/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 15:58:34 | examples-with-cassandra/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 15:58:35 | examples-with-cassandra/2-smoke-test | job.batch/report-span created logger.go:42: 15:58:35 | examples-with-cassandra/2-smoke-test | job.batch/check-span created logger.go:42: 15:58:46 | examples-with-cassandra/2-smoke-test | test step completed 2-smoke-test logger.go:42: 15:58:46 | examples-with-cassandra | examples-with-cassandra events from ns kuttl-test-funny-airedale: logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:57:45 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-funny-airedale/cassandra-0 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:57:45 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.129.2.30/23] from ovn-kubernetes multus logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:57:45 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:57:45 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:57:46 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:57:46 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:57:48 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-funny-airedale/cassandra-1 to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:57:48 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.131.0.57/23] from ovn-kubernetes multus logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:57:48 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:57:48 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:57:48 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:57:48 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:57:49 +0000 UTC Warning Pod cassandra-0.spec.containers{cassandra} BackOff Back-off restarting failed container cassandra in pod cassandra-0_kuttl-test-funny-airedale(42b6111a-fd63-420c-b28d-32cf49c5f70a) kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:57:52 +0000 UTC Warning Pod cassandra-1.spec.containers{cassandra} BackOff Back-off restarting failed container cassandra in pod cassandra-1_kuttl-test-funny-airedale(da6c7083-10d8-4f63-b954-4eefafb26278) kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:12 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-jkl5g Binding Scheduled Successfully assigned kuttl-test-funny-airedale/with-cassandra-cassandra-schema-job-jkl5g to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:12 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-jkl5g AddedInterface Add eth0 [10.129.2.31/23] from ovn-kubernetes multus logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:12 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job SuccessfulCreate Created pod: with-cassandra-cassandra-schema-job-jkl5g job-controller logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:13 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-jkl5g.spec.containers{with-cassandra-cassandra-schema-job} Pulling Pulling image "jaegertracing/jaeger-cassandra-schema:1.57.0" kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:18 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-jkl5g.spec.containers{with-cassandra-cassandra-schema-job} Pulled Successfully pulled image "jaegertracing/jaeger-cassandra-schema:1.57.0" in 5.204s (5.204s including waiting) kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:18 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-jkl5g.spec.containers{with-cassandra-cassandra-schema-job} Created Created container with-cassandra-cassandra-schema-job kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:18 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-jkl5g.spec.containers{with-cassandra-cassandra-schema-job} Started Started container with-cassandra-cassandra-schema-job kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:23 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job Completed Job completed job-controller logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:23 +0000 UTC Normal Pod with-cassandra-fff8f46b8-n6sct Binding Scheduled Successfully assigned kuttl-test-funny-airedale/with-cassandra-fff8f46b8-n6sct to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:23 +0000 UTC Normal ReplicaSet.apps with-cassandra-fff8f46b8 SuccessfulCreate Created pod: with-cassandra-fff8f46b8-n6sct replicaset-controller logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:23 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-fff8f46b8 to 1 deployment-controller logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:24 +0000 UTC Normal Pod with-cassandra-fff8f46b8-n6sct AddedInterface Add eth0 [10.129.2.32/23] from ovn-kubernetes multus logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:24 +0000 UTC Normal Pod with-cassandra-fff8f46b8-n6sct.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:24 +0000 UTC Normal Pod with-cassandra-fff8f46b8-n6sct.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:24 +0000 UTC Normal Pod with-cassandra-fff8f46b8-n6sct.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:24 +0000 UTC Normal Pod with-cassandra-fff8f46b8-n6sct.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:24 +0000 UTC Normal Pod with-cassandra-fff8f46b8-n6sct.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:24 +0000 UTC Normal Pod with-cassandra-fff8f46b8-n6sct.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:30 +0000 UTC Normal Pod with-cassandra-fff8f46b8-n6sct.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:30 +0000 UTC Normal Pod with-cassandra-fff8f46b8-n6sct.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:30 +0000 UTC Normal ReplicaSet.apps with-cassandra-fff8f46b8 SuccessfulDelete Deleted pod: with-cassandra-fff8f46b8-n6sct replicaset-controller logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:30 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled down replica set with-cassandra-fff8f46b8 to 0 from 1 deployment-controller logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:31 +0000 UTC Normal Pod with-cassandra-55cc847479-nplp4 Binding Scheduled Successfully assigned kuttl-test-funny-airedale/with-cassandra-55cc847479-nplp4 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:31 +0000 UTC Normal Pod with-cassandra-55cc847479-nplp4 AddedInterface Add eth0 [10.129.2.33/23] from ovn-kubernetes multus logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:31 +0000 UTC Normal Pod with-cassandra-55cc847479-nplp4.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:31 +0000 UTC Normal Pod with-cassandra-55cc847479-nplp4.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:31 +0000 UTC Normal ReplicaSet.apps with-cassandra-55cc847479 SuccessfulCreate Created pod: with-cassandra-55cc847479-nplp4 replicaset-controller logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:31 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-55cc847479 to 1 deployment-controller logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:32 +0000 UTC Normal Pod with-cassandra-55cc847479-nplp4.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:32 +0000 UTC Normal Pod with-cassandra-55cc847479-nplp4.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:32 +0000 UTC Normal Pod with-cassandra-55cc847479-nplp4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:32 +0000 UTC Normal Pod with-cassandra-55cc847479-nplp4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:35 +0000 UTC Normal Pod check-span-plxg6 Binding Scheduled Successfully assigned kuttl-test-funny-airedale/check-span-plxg6 to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:35 +0000 UTC Normal Pod check-span-plxg6 AddedInterface Add eth0 [10.131.0.58/23] from ovn-kubernetes multus logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:35 +0000 UTC Normal Pod check-span-plxg6.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:35 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-plxg6 job-controller logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:35 +0000 UTC Normal Pod report-span-svsz7 Binding Scheduled Successfully assigned kuttl-test-funny-airedale/report-span-svsz7 to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:35 +0000 UTC Normal Pod report-span-svsz7 AddedInterface Add eth0 [10.128.2.34/23] from ovn-kubernetes multus logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:35 +0000 UTC Normal Pod report-span-svsz7.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:35 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-svsz7 job-controller logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:36 +0000 UTC Normal Pod check-span-plxg6.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 331ms (331ms including waiting) kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:36 +0000 UTC Normal Pod check-span-plxg6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:36 +0000 UTC Normal Pod check-span-plxg6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:36 +0000 UTC Normal Pod report-span-svsz7.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 378ms (378ms including waiting) kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:36 +0000 UTC Normal Pod report-span-svsz7.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:36 +0000 UTC Normal Pod report-span-svsz7.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 15:58:46 | examples-with-cassandra | 2024-06-17 15:58:46 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 15:58:46 | examples-with-cassandra | Deleting namespace: kuttl-test-funny-airedale === CONT kuttl/harness/examples-with-badger logger.go:42: 15:58:59 | examples-with-badger | Creating namespace: kuttl-test-sought-elf logger.go:42: 15:58:59 | examples-with-badger/0-install | starting test step 0-install logger.go:42: 15:58:59 | examples-with-badger/0-install | Jaeger:kuttl-test-sought-elf/with-badger created logger.go:42: 15:59:04 | examples-with-badger/0-install | test step completed 0-install logger.go:42: 15:59:04 | examples-with-badger/1-smoke-test | starting test step 1-smoke-test logger.go:42: 15:59:04 | examples-with-badger/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger /dev/null] logger.go:42: 15:59:06 | examples-with-badger/1-smoke-test | Warning: resource jaegers/with-badger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 15:59:13 | examples-with-badger/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 15:59:13 | examples-with-badger/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 15:59:13 | examples-with-badger/1-smoke-test | job.batch/report-span created logger.go:42: 15:59:14 | examples-with-badger/1-smoke-test | job.batch/check-span created logger.go:42: 15:59:26 | examples-with-badger/1-smoke-test | test step completed 1-smoke-test logger.go:42: 15:59:26 | examples-with-badger | examples-with-badger events from ns kuttl-test-sought-elf: logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:03 +0000 UTC Normal Pod with-badger-6d8799bff8-fkdxs Binding Scheduled Successfully assigned kuttl-test-sought-elf/with-badger-6d8799bff8-fkdxs to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:03 +0000 UTC Normal Pod with-badger-6d8799bff8-fkdxs AddedInterface Add eth0 [10.129.2.34/23] from ovn-kubernetes multus logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:03 +0000 UTC Normal Pod with-badger-6d8799bff8-fkdxs.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:03 +0000 UTC Normal Pod with-badger-6d8799bff8-fkdxs.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:03 +0000 UTC Normal Pod with-badger-6d8799bff8-fkdxs.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:03 +0000 UTC Normal Pod with-badger-6d8799bff8-fkdxs.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:03 +0000 UTC Normal ReplicaSet.apps with-badger-6d8799bff8 SuccessfulCreate Created pod: with-badger-6d8799bff8-fkdxs replicaset-controller logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:03 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-6d8799bff8 to 1 deployment-controller logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:04 +0000 UTC Normal Pod with-badger-6d8799bff8-fkdxs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:04 +0000 UTC Normal Pod with-badger-6d8799bff8-fkdxs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:08 +0000 UTC Normal Pod with-badger-6d8799bff8-fkdxs.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:08 +0000 UTC Normal Pod with-badger-6d8799bff8-fkdxs.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:08 +0000 UTC Normal ReplicaSet.apps with-badger-6d8799bff8 SuccessfulDelete Deleted pod: with-badger-6d8799bff8-fkdxs replicaset-controller logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:08 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled down replica set with-badger-6d8799bff8 to 0 from 1 deployment-controller logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:09 +0000 UTC Normal Pod with-badger-5f77988bc7-qqnpc Binding Scheduled Successfully assigned kuttl-test-sought-elf/with-badger-5f77988bc7-qqnpc to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:09 +0000 UTC Normal ReplicaSet.apps with-badger-5f77988bc7 SuccessfulCreate Created pod: with-badger-5f77988bc7-qqnpc replicaset-controller logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:09 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-5f77988bc7 to 1 deployment-controller logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:10 +0000 UTC Normal Pod with-badger-5f77988bc7-qqnpc AddedInterface Add eth0 [10.129.2.35/23] from ovn-kubernetes multus logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:10 +0000 UTC Normal Pod with-badger-5f77988bc7-qqnpc.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:10 +0000 UTC Normal Pod with-badger-5f77988bc7-qqnpc.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:10 +0000 UTC Normal Pod with-badger-5f77988bc7-qqnpc.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:10 +0000 UTC Normal Pod with-badger-5f77988bc7-qqnpc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:10 +0000 UTC Normal Pod with-badger-5f77988bc7-qqnpc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:10 +0000 UTC Normal Pod with-badger-5f77988bc7-qqnpc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:13 +0000 UTC Normal Pod report-span-cpt6r Binding Scheduled Successfully assigned kuttl-test-sought-elf/report-span-cpt6r to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:13 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-cpt6r job-controller logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:14 +0000 UTC Normal Pod check-span-pqp4n Binding Scheduled Successfully assigned kuttl-test-sought-elf/check-span-pqp4n to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:14 +0000 UTC Normal Pod check-span-pqp4n AddedInterface Add eth0 [10.128.2.35/23] from ovn-kubernetes multus logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:14 +0000 UTC Normal Pod check-span-pqp4n.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:14 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-pqp4n job-controller logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:14 +0000 UTC Normal Pod report-span-cpt6r AddedInterface Add eth0 [10.131.0.59/23] from ovn-kubernetes multus logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:14 +0000 UTC Normal Pod report-span-cpt6r.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:14 +0000 UTC Normal Pod report-span-cpt6r.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 321ms (321ms including waiting) kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:15 +0000 UTC Normal Pod check-span-pqp4n.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 330ms (330ms including waiting) kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:15 +0000 UTC Normal Pod check-span-pqp4n.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:15 +0000 UTC Normal Pod check-span-pqp4n.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:15 +0000 UTC Normal Pod report-span-cpt6r.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:15 +0000 UTC Normal Pod report-span-cpt6r.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 15:59:26 | examples-with-badger | 2024-06-17 15:59:25 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 15:59:26 | examples-with-badger | Deleting namespace: kuttl-test-sought-elf === CONT kuttl/harness/examples-simplest logger.go:42: 15:59:34 | examples-simplest | Creating namespace: kuttl-test-eager-mole logger.go:42: 15:59:34 | examples-simplest/0-install | starting test step 0-install logger.go:42: 15:59:35 | examples-simplest/0-install | Jaeger:kuttl-test-eager-mole/simplest created logger.go:42: 15:59:41 | examples-simplest/0-install | test step completed 0-install logger.go:42: 15:59:41 | examples-simplest/1-smoke-test | starting test step 1-smoke-test logger.go:42: 15:59:41 | examples-simplest/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 15:59:42 | examples-simplest/1-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 15:59:49 | examples-simplest/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 15:59:49 | examples-simplest/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 15:59:50 | examples-simplest/1-smoke-test | job.batch/report-span created logger.go:42: 15:59:50 | examples-simplest/1-smoke-test | job.batch/check-span created logger.go:42: 16:00:02 | examples-simplest/1-smoke-test | test step completed 1-smoke-test logger.go:42: 16:00:02 | examples-simplest | examples-simplest events from ns kuttl-test-eager-mole: logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:38 +0000 UTC Normal Pod simplest-6dfb5ffb45-ktz52 Binding Scheduled Successfully assigned kuttl-test-eager-mole/simplest-6dfb5ffb45-ktz52 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:38 +0000 UTC Normal ReplicaSet.apps simplest-6dfb5ffb45 SuccessfulCreate Created pod: simplest-6dfb5ffb45-ktz52 replicaset-controller logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:38 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-6dfb5ffb45 to 1 deployment-controller logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:39 +0000 UTC Normal Pod simplest-6dfb5ffb45-ktz52 AddedInterface Add eth0 [10.129.2.36/23] from ovn-kubernetes multus logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:39 +0000 UTC Normal Pod simplest-6dfb5ffb45-ktz52.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:39 +0000 UTC Normal Pod simplest-6dfb5ffb45-ktz52.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:39 +0000 UTC Normal Pod simplest-6dfb5ffb45-ktz52.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:39 +0000 UTC Normal Pod simplest-6dfb5ffb45-ktz52.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:39 +0000 UTC Normal Pod simplest-6dfb5ffb45-ktz52.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:39 +0000 UTC Normal Pod simplest-6dfb5ffb45-ktz52.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:44 +0000 UTC Normal Pod simplest-6dfb5ffb45-ktz52.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:44 +0000 UTC Normal Pod simplest-6dfb5ffb45-ktz52.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:44 +0000 UTC Normal ReplicaSet.apps simplest-6dfb5ffb45 SuccessfulDelete Deleted pod: simplest-6dfb5ffb45-ktz52 replicaset-controller logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:44 +0000 UTC Normal Pod simplest-7cb9d4f796-m6nf6 Binding Scheduled Successfully assigned kuttl-test-eager-mole/simplest-7cb9d4f796-m6nf6 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:44 +0000 UTC Normal ReplicaSet.apps simplest-7cb9d4f796 SuccessfulCreate Created pod: simplest-7cb9d4f796-m6nf6 replicaset-controller logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:44 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-6dfb5ffb45 to 0 from 1 deployment-controller logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:44 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-7cb9d4f796 to 1 deployment-controller logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:45 +0000 UTC Normal Pod simplest-7cb9d4f796-m6nf6 AddedInterface Add eth0 [10.129.2.37/23] from ovn-kubernetes multus logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:45 +0000 UTC Normal Pod simplest-7cb9d4f796-m6nf6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:45 +0000 UTC Normal Pod simplest-7cb9d4f796-m6nf6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:45 +0000 UTC Normal Pod simplest-7cb9d4f796-m6nf6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:45 +0000 UTC Normal Pod simplest-7cb9d4f796-m6nf6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:45 +0000 UTC Normal Pod simplest-7cb9d4f796-m6nf6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:45 +0000 UTC Normal Pod simplest-7cb9d4f796-m6nf6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:50 +0000 UTC Normal Pod check-span-9hph4 Binding Scheduled Successfully assigned kuttl-test-eager-mole/check-span-9hph4 to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:50 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-9hph4 job-controller logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:50 +0000 UTC Normal Pod report-span-4xggd Binding Scheduled Successfully assigned kuttl-test-eager-mole/report-span-4xggd to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:50 +0000 UTC Normal Pod report-span-4xggd AddedInterface Add eth0 [10.131.0.60/23] from ovn-kubernetes multus logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:50 +0000 UTC Normal Pod report-span-4xggd.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:50 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-4xggd job-controller logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:51 +0000 UTC Normal Pod check-span-9hph4 AddedInterface Add eth0 [10.128.2.36/23] from ovn-kubernetes multus logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:51 +0000 UTC Normal Pod check-span-9hph4.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:51 +0000 UTC Normal Pod check-span-9hph4.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 347ms (347ms including waiting) kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:51 +0000 UTC Normal Pod check-span-9hph4.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:51 +0000 UTC Normal Pod check-span-9hph4.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:51 +0000 UTC Normal Pod report-span-4xggd.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 360ms (360ms including waiting) kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:51 +0000 UTC Normal Pod report-span-4xggd.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 15:59:51 +0000 UTC Normal Pod report-span-4xggd.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 16:00:02 | examples-simplest | 2024-06-17 16:00:02 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 16:00:02 | examples-simplest | Deleting namespace: kuttl-test-eager-mole === CONT kuttl/harness/examples-simple-prod-with-volumes logger.go:42: 16:00:15 | examples-simple-prod-with-volumes | Ignoring 03-check-volume.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 16:00:15 | examples-simple-prod-with-volumes | Creating namespace: kuttl-test-meet-meerkat logger.go:42: 16:00:15 | examples-simple-prod-with-volumes/1-install | starting test step 1-install logger.go:42: 16:00:15 | examples-simple-prod-with-volumes/1-install | Jaeger:kuttl-test-meet-meerkat/simple-prod created logger.go:42: 16:00:51 | examples-simple-prod-with-volumes/1-install | test step completed 1-install logger.go:42: 16:00:51 | examples-simple-prod-with-volumes/2-smoke-test | starting test step 2-smoke-test logger.go:42: 16:00:51 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 16:00:53 | examples-simple-prod-with-volumes/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 16:00:59 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 16:01:00 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 16:01:00 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/report-span created logger.go:42: 16:01:00 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/check-span created logger.go:42: 16:01:13 | examples-simple-prod-with-volumes/2-smoke-test | test step completed 2-smoke-test logger.go:42: 16:01:13 | examples-simple-prod-with-volumes/3-check-volume | starting test step 3-check-volume logger.go:42: 16:01:13 | examples-simple-prod-with-volumes/3-check-volume | running command: [sh -c kubectl exec $(kubectl get pods -n $NAMESPACE -l app=jaeger -l app.kubernetes.io/component=collector -o yaml | /tmp/jaeger-tests/bin/yq e '.items[0].metadata.name') -n $NAMESPACE -- ls /usr/share/elasticsearch/data] logger.go:42: 16:01:13 | examples-simple-prod-with-volumes/3-check-volume | test step completed 3-check-volume logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | examples-simple-prod-with-volumes events from ns kuttl-test-meet-meerkat: logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:20 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestmeetmeerkatsimpleprod-1-7f7fbd879b SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestmeetmeerkatsimpleprod-1-7f7fbd8f9gh7 replicaset-controller logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeetmeerkatsimpleprod-1-7f7fbd8f9gh7 Binding Scheduled Successfully assigned kuttl-test-meet-meerkat/elasticsearch-cdm-kuttltestmeetmeerkatsimpleprod-1-7f7fbd8f9gh7 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:20 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestmeetmeerkatsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestmeetmeerkatsimpleprod-1-7f7fbd879b to 1 deployment-controller logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeetmeerkatsimpleprod-1-7f7fbd8f9gh7 AddedInterface Add eth0 [10.129.2.38/23] from ovn-kubernetes multus logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeetmeerkatsimpleprod-1-7f7fbd8f9gh7.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeetmeerkatsimpleprod-1-7f7fbd8f9gh7.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeetmeerkatsimpleprod-1-7f7fbd8f9gh7.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeetmeerkatsimpleprod-1-7f7fbd8f9gh7.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeetmeerkatsimpleprod-1-7f7fbd8f9gh7.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeetmeerkatsimpleprod-1-7f7fbd8f9gh7.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:31 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmeetmeerkatsimpleprod-1-7f7fbd8f9gh7.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:36 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmeetmeerkatsimpleprod-1-7f7fbd8f9gh7.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:47 +0000 UTC Normal Pod simple-prod-collector-bdbdb9848-7j6qc Binding Scheduled Successfully assigned kuttl-test-meet-meerkat/simple-prod-collector-bdbdb9848-7j6qc to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:47 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-bdbdb9848 SuccessfulCreate Created pod: simple-prod-collector-bdbdb9848-7j6qc replicaset-controller logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:47 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-bdbdb9848 to 1 deployment-controller logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:47 +0000 UTC Normal Pod simple-prod-query-5f7dbc5f78-xmflr Binding Scheduled Successfully assigned kuttl-test-meet-meerkat/simple-prod-query-5f7dbc5f78-xmflr to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:47 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5f7dbc5f78 SuccessfulCreate Created pod: simple-prod-query-5f7dbc5f78-xmflr replicaset-controller logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:47 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-5f7dbc5f78 to 1 deployment-controller logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:48 +0000 UTC Normal Pod simple-prod-collector-bdbdb9848-7j6qc AddedInterface Add eth0 [10.131.0.62/23] from ovn-kubernetes multus logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:48 +0000 UTC Normal Pod simple-prod-collector-bdbdb9848-7j6qc.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1" already present on machine kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:48 +0000 UTC Normal Pod simple-prod-collector-bdbdb9848-7j6qc.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:48 +0000 UTC Normal Pod simple-prod-collector-bdbdb9848-7j6qc.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:49 +0000 UTC Warning Pod simple-prod-query-5f7dbc5f78-xmflr FailedMount MountVolume.SetUp failed for volume "kube-api-access-djncc" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:50 +0000 UTC Normal Pod simple-prod-query-5f7dbc5f78-xmflr AddedInterface Add eth0 [10.128.2.37/23] from ovn-kubernetes multus logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:50 +0000 UTC Normal Pod simple-prod-query-5f7dbc5f78-xmflr.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:50 +0000 UTC Normal Pod simple-prod-query-5f7dbc5f78-xmflr.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:50 +0000 UTC Normal Pod simple-prod-query-5f7dbc5f78-xmflr.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:50 +0000 UTC Normal Pod simple-prod-query-5f7dbc5f78-xmflr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:50 +0000 UTC Normal Pod simple-prod-query-5f7dbc5f78-xmflr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:50 +0000 UTC Normal Pod simple-prod-query-5f7dbc5f78-xmflr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:50 +0000 UTC Normal Pod simple-prod-query-5f7dbc5f78-xmflr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:50 +0000 UTC Normal Pod simple-prod-query-5f7dbc5f78-xmflr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:50 +0000 UTC Normal Pod simple-prod-query-5f7dbc5f78-xmflr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:54 +0000 UTC Normal Pod simple-prod-query-5f7dbc5f78-xmflr.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:54 +0000 UTC Normal Pod simple-prod-query-5f7dbc5f78-xmflr.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:54 +0000 UTC Normal Pod simple-prod-query-5f7dbc5f78-xmflr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:54 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5f7dbc5f78 SuccessfulDelete Deleted pod: simple-prod-query-5f7dbc5f78-xmflr replicaset-controller logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:54 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-5f7dbc5f78 to 0 from 1 deployment-controller logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:55 +0000 UTC Normal Pod simple-prod-query-6ff878d76f-gtmg5 Binding Scheduled Successfully assigned kuttl-test-meet-meerkat/simple-prod-query-6ff878d76f-gtmg5 to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:55 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6ff878d76f SuccessfulCreate Created pod: simple-prod-query-6ff878d76f-gtmg5 replicaset-controller logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:55 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6ff878d76f to 1 deployment-controller logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:56 +0000 UTC Normal Pod simple-prod-query-6ff878d76f-gtmg5 AddedInterface Add eth0 [10.128.2.38/23] from ovn-kubernetes multus logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:56 +0000 UTC Normal Pod simple-prod-query-6ff878d76f-gtmg5.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:56 +0000 UTC Normal Pod simple-prod-query-6ff878d76f-gtmg5.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:56 +0000 UTC Normal Pod simple-prod-query-6ff878d76f-gtmg5.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:56 +0000 UTC Normal Pod simple-prod-query-6ff878d76f-gtmg5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:56 +0000 UTC Normal Pod simple-prod-query-6ff878d76f-gtmg5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:56 +0000 UTC Normal Pod simple-prod-query-6ff878d76f-gtmg5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:56 +0000 UTC Normal Pod simple-prod-query-6ff878d76f-gtmg5.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:56 +0000 UTC Normal Pod simple-prod-query-6ff878d76f-gtmg5.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:00:56 +0000 UTC Normal Pod simple-prod-query-6ff878d76f-gtmg5.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:01:00 +0000 UTC Normal Pod check-span-d9btn Binding Scheduled Successfully assigned kuttl-test-meet-meerkat/check-span-d9btn to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:01:00 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-d9btn job-controller logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:01:00 +0000 UTC Normal Pod report-span-7nf2v Binding Scheduled Successfully assigned kuttl-test-meet-meerkat/report-span-7nf2v to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:01:00 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-7nf2v job-controller logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:01:01 +0000 UTC Normal Pod check-span-d9btn AddedInterface Add eth0 [10.131.0.64/23] from ovn-kubernetes multus logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:01:01 +0000 UTC Normal Pod check-span-d9btn.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:01:01 +0000 UTC Normal Pod check-span-d9btn.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 376ms (376ms including waiting) kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:01:01 +0000 UTC Normal Pod report-span-7nf2v AddedInterface Add eth0 [10.131.0.63/23] from ovn-kubernetes multus logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:01:01 +0000 UTC Normal Pod report-span-7nf2v.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:01:01 +0000 UTC Normal Pod report-span-7nf2v.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 447ms (447ms including waiting) kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:01:02 +0000 UTC Normal Pod check-span-d9btn.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:01:02 +0000 UTC Normal Pod check-span-d9btn.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:01:02 +0000 UTC Normal Pod report-span-7nf2v.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:01:02 +0000 UTC Normal Pod report-span-7nf2v.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:01:02 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:01:02 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:01:02 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | 2024-06-17 16:01:12 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 16:01:13 | examples-simple-prod-with-volumes | Deleting namespace: kuttl-test-meet-meerkat === CONT kuttl/harness/examples-business-application-injected-sidecar logger.go:42: 16:01:26 | examples-business-application-injected-sidecar | Creating namespace: kuttl-test-finer-toad logger.go:42: 16:01:26 | examples-business-application-injected-sidecar/0-install | starting test step 0-install logger.go:42: 16:01:26 | examples-business-application-injected-sidecar/0-install | Deployment:kuttl-test-finer-toad/myapp created logger.go:42: 16:01:26 | examples-business-application-injected-sidecar/0-install | test step completed 0-install logger.go:42: 16:01:26 | examples-business-application-injected-sidecar/1-install | starting test step 1-install logger.go:42: 16:01:26 | examples-business-application-injected-sidecar/1-install | Jaeger:kuttl-test-finer-toad/simplest created logger.go:42: 16:01:38 | examples-business-application-injected-sidecar/1-install | test step completed 1-install logger.go:42: 16:01:38 | examples-business-application-injected-sidecar/2-smoke-test | starting test step 2-smoke-test logger.go:42: 16:01:38 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 16:01:39 | examples-business-application-injected-sidecar/2-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 16:01:46 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 16:01:47 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 16:01:47 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/report-span created logger.go:42: 16:01:47 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/check-span created logger.go:42: 16:01:53 | examples-business-application-injected-sidecar/2-smoke-test | test step completed 2-smoke-test logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | examples-business-application-injected-sidecar events from ns kuttl-test-finer-toad: logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:26 +0000 UTC Normal Pod myapp-679f79d5f8-8wj95 Binding Scheduled Successfully assigned kuttl-test-finer-toad/myapp-679f79d5f8-8wj95 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:26 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulCreate Created pod: myapp-679f79d5f8-8wj95 replicaset-controller logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:26 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-679f79d5f8 to 1 deployment-controller logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:27 +0000 UTC Normal Pod myapp-679f79d5f8-8wj95 AddedInterface Add eth0 [10.129.2.39/23] from ovn-kubernetes multus logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:27 +0000 UTC Normal Pod myapp-679f79d5f8-8wj95.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:27 +0000 UTC Normal Pod myapp-6d7d9c6cff-bxgzr Binding Scheduled Successfully assigned kuttl-test-finer-toad/myapp-6d7d9c6cff-bxgzr to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:27 +0000 UTC Normal ReplicaSet.apps myapp-6d7d9c6cff SuccessfulCreate Created pod: myapp-6d7d9c6cff-bxgzr replicaset-controller logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:27 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-6d7d9c6cff to 1 deployment-controller logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:28 +0000 UTC Warning Pod myapp-6d7d9c6cff-bxgzr FailedMount MountVolume.SetUp failed for volume "simplest-service-ca" : configmap "simplest-service-ca" not found kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:28 +0000 UTC Warning Pod myapp-6d7d9c6cff-bxgzr FailedMount MountVolume.SetUp failed for volume "simplest-trusted-ca" : configmap "simplest-trusted-ca" not found kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:31 +0000 UTC Normal Pod myapp-679f79d5f8-8wj95.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.98s (3.98s including waiting) kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:31 +0000 UTC Normal Pod myapp-679f79d5f8-8wj95.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:31 +0000 UTC Normal Pod myapp-679f79d5f8-8wj95.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:35 +0000 UTC Normal Pod simplest-65fd85bbb5-w5qww Binding Scheduled Successfully assigned kuttl-test-finer-toad/simplest-65fd85bbb5-w5qww to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:35 +0000 UTC Normal ReplicaSet.apps simplest-65fd85bbb5 SuccessfulCreate Created pod: simplest-65fd85bbb5-w5qww replicaset-controller logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:35 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-65fd85bbb5 to 1 deployment-controller logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:36 +0000 UTC Normal Pod myapp-6d7d9c6cff-bxgzr AddedInterface Add eth0 [10.131.0.65/23] from ovn-kubernetes multus logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:36 +0000 UTC Normal Pod myapp-6d7d9c6cff-bxgzr.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:36 +0000 UTC Normal Pod simplest-65fd85bbb5-w5qww AddedInterface Add eth0 [10.129.2.40/23] from ovn-kubernetes multus logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:36 +0000 UTC Normal Pod simplest-65fd85bbb5-w5qww.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:36 +0000 UTC Normal Pod simplest-65fd85bbb5-w5qww.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:36 +0000 UTC Normal Pod simplest-65fd85bbb5-w5qww.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:36 +0000 UTC Normal Pod simplest-65fd85bbb5-w5qww.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:36 +0000 UTC Normal Pod simplest-65fd85bbb5-w5qww.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:36 +0000 UTC Normal Pod simplest-65fd85bbb5-w5qww.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:38 +0000 UTC Warning Pod myapp-679f79d5f8-8wj95.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.129.2.39:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:40 +0000 UTC Normal Pod myapp-6d7d9c6cff-bxgzr.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 4.363s (4.363s including waiting) kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:40 +0000 UTC Normal Pod myapp-6d7d9c6cff-bxgzr.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:40 +0000 UTC Normal Pod myapp-6d7d9c6cff-bxgzr.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:40 +0000 UTC Normal Pod myapp-6d7d9c6cff-bxgzr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:41 +0000 UTC Normal Pod myapp-6d7d9c6cff-bxgzr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:41 +0000 UTC Normal Pod myapp-6d7d9c6cff-bxgzr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:41 +0000 UTC Normal Pod simplest-65fd85bbb5-w5qww.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:41 +0000 UTC Normal Pod simplest-65fd85bbb5-w5qww.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:41 +0000 UTC Normal ReplicaSet.apps simplest-65fd85bbb5 SuccessfulDelete Deleted pod: simplest-65fd85bbb5-w5qww replicaset-controller logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:41 +0000 UTC Normal Pod simplest-7bb45cdb85-7vl5x Binding Scheduled Successfully assigned kuttl-test-finer-toad/simplest-7bb45cdb85-7vl5x to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:41 +0000 UTC Normal ReplicaSet.apps simplest-7bb45cdb85 SuccessfulCreate Created pod: simplest-7bb45cdb85-7vl5x replicaset-controller logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:41 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-65fd85bbb5 to 0 from 1 deployment-controller logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:41 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-7bb45cdb85 to 1 deployment-controller logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:42 +0000 UTC Normal Pod myapp-679f79d5f8-8wj95.spec.containers{myapp} Killing Stopping container myapp kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:42 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulDelete Deleted pod: myapp-679f79d5f8-8wj95 replicaset-controller logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:42 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled down replica set myapp-679f79d5f8 to 0 from 1 deployment-controller logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:42 +0000 UTC Normal Pod simplest-7bb45cdb85-7vl5x AddedInterface Add eth0 [10.129.2.41/23] from ovn-kubernetes multus logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:42 +0000 UTC Normal Pod simplest-7bb45cdb85-7vl5x.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:42 +0000 UTC Normal Pod simplest-7bb45cdb85-7vl5x.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:42 +0000 UTC Normal Pod simplest-7bb45cdb85-7vl5x.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:42 +0000 UTC Normal Pod simplest-7bb45cdb85-7vl5x.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:42 +0000 UTC Normal Pod simplest-7bb45cdb85-7vl5x.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:42 +0000 UTC Normal Pod simplest-7bb45cdb85-7vl5x.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:47 +0000 UTC Normal Pod check-span-nwtf7 Binding Scheduled Successfully assigned kuttl-test-finer-toad/check-span-nwtf7 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:47 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-nwtf7 job-controller logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:47 +0000 UTC Warning Pod myapp-6d7d9c6cff-bxgzr.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.131.0.65:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:47 +0000 UTC Normal Pod report-span-25rj7 Binding Scheduled Successfully assigned kuttl-test-finer-toad/report-span-25rj7 to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:47 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-25rj7 job-controller logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:48 +0000 UTC Normal Pod check-span-nwtf7 AddedInterface Add eth0 [10.129.2.42/23] from ovn-kubernetes multus logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:48 +0000 UTC Normal Pod check-span-nwtf7.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:48 +0000 UTC Normal Pod report-span-25rj7 AddedInterface Add eth0 [10.128.2.39/23] from ovn-kubernetes multus logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:48 +0000 UTC Normal Pod report-span-25rj7.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:48 +0000 UTC Normal Pod report-span-25rj7.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 449ms (449ms including waiting) kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:48 +0000 UTC Normal Pod report-span-25rj7.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:48 +0000 UTC Normal Pod report-span-25rj7.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:50 +0000 UTC Normal Pod check-span-nwtf7.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.915s (1.915s including waiting) kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:50 +0000 UTC Normal Pod check-span-nwtf7.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:50 +0000 UTC Normal Pod check-span-nwtf7.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | 2024-06-17 16:01:52 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 16:01:53 | examples-business-application-injected-sidecar | Deleting namespace: kuttl-test-finer-toad === CONT kuttl/harness/examples-service-types logger.go:42: 16:02:07 | examples-service-types | Creating namespace: kuttl-test-thankful-akita logger.go:42: 16:02:07 | examples-service-types/0-install | starting test step 0-install logger.go:42: 16:02:07 | examples-service-types/0-install | Jaeger:kuttl-test-thankful-akita/service-types created logger.go:42: 16:02:15 | examples-service-types/0-install | test step completed 0-install logger.go:42: 16:02:15 | examples-service-types/1-smoke-test | starting test step 1-smoke-test logger.go:42: 16:02:15 | examples-service-types/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE service-types /dev/null] logger.go:42: 16:02:17 | examples-service-types/1-smoke-test | Warning: resource jaegers/service-types is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 16:02:23 | examples-service-types/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://service-types-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 16:02:24 | examples-service-types/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 16:02:24 | examples-service-types/1-smoke-test | job.batch/report-span created logger.go:42: 16:02:24 | examples-service-types/1-smoke-test | job.batch/check-span created logger.go:42: 16:02:37 | examples-service-types/1-smoke-test | test step completed 1-smoke-test logger.go:42: 16:02:37 | examples-service-types/2- | starting test step 2- logger.go:42: 16:02:37 | examples-service-types/2- | test step completed 2- logger.go:42: 16:02:37 | examples-service-types | examples-service-types events from ns kuttl-test-thankful-akita: logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:11 +0000 UTC Normal ReplicaSet.apps service-types-5f98f74b95 SuccessfulCreate Created pod: service-types-5f98f74b95-pzsd9 replicaset-controller logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:11 +0000 UTC Normal Service service-types-collector EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:11 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-5f98f74b95 to 1 deployment-controller logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:12 +0000 UTC Normal Pod service-types-5f98f74b95-pzsd9 Binding Scheduled Successfully assigned kuttl-test-thankful-akita/service-types-5f98f74b95-pzsd9 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:12 +0000 UTC Normal Pod service-types-5f98f74b95-pzsd9 AddedInterface Add eth0 [10.129.2.43/23] from ovn-kubernetes multus logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:12 +0000 UTC Normal Pod service-types-5f98f74b95-pzsd9.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:12 +0000 UTC Normal Pod service-types-5f98f74b95-pzsd9.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:12 +0000 UTC Normal Pod service-types-5f98f74b95-pzsd9.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:12 +0000 UTC Normal Pod service-types-5f98f74b95-pzsd9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:12 +0000 UTC Normal Pod service-types-5f98f74b95-pzsd9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:12 +0000 UTC Normal Pod service-types-5f98f74b95-pzsd9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:14 +0000 UTC Normal Service service-types-collector EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:15 +0000 UTC Normal Service service-types-query EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:17 +0000 UTC Normal Service service-types-query EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:18 +0000 UTC Normal Pod service-types-5f98f74b95-pzsd9.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:18 +0000 UTC Normal Pod service-types-5f98f74b95-pzsd9.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:18 +0000 UTC Normal ReplicaSet.apps service-types-5f98f74b95 SuccessfulDelete Deleted pod: service-types-5f98f74b95-pzsd9 replicaset-controller logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:18 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled down replica set service-types-5f98f74b95 to 0 from 1 deployment-controller logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:19 +0000 UTC Normal Pod service-types-f5487b45d-tfpdh Binding Scheduled Successfully assigned kuttl-test-thankful-akita/service-types-f5487b45d-tfpdh to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:19 +0000 UTC Normal ReplicaSet.apps service-types-f5487b45d SuccessfulCreate Created pod: service-types-f5487b45d-tfpdh replicaset-controller logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:19 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-f5487b45d to 1 deployment-controller logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:20 +0000 UTC Normal Pod service-types-f5487b45d-tfpdh AddedInterface Add eth0 [10.129.2.44/23] from ovn-kubernetes multus logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:20 +0000 UTC Normal Pod service-types-f5487b45d-tfpdh.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:20 +0000 UTC Normal Pod service-types-f5487b45d-tfpdh.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:20 +0000 UTC Normal Pod service-types-f5487b45d-tfpdh.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:20 +0000 UTC Normal Pod service-types-f5487b45d-tfpdh.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:20 +0000 UTC Normal Pod service-types-f5487b45d-tfpdh.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:20 +0000 UTC Normal Pod service-types-f5487b45d-tfpdh.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:24 +0000 UTC Normal Pod check-span-6nr7s Binding Scheduled Successfully assigned kuttl-test-thankful-akita/check-span-6nr7s to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:24 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-6nr7s job-controller logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:24 +0000 UTC Normal Pod report-span-kbzjj Binding Scheduled Successfully assigned kuttl-test-thankful-akita/report-span-kbzjj to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:24 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-kbzjj job-controller logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:25 +0000 UTC Normal Pod check-span-6nr7s AddedInterface Add eth0 [10.128.2.40/23] from ovn-kubernetes multus logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:25 +0000 UTC Normal Pod check-span-6nr7s.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:25 +0000 UTC Normal Pod check-span-6nr7s.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 406ms (406ms including waiting) kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:25 +0000 UTC Normal Pod check-span-6nr7s.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:25 +0000 UTC Normal Pod check-span-6nr7s.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:25 +0000 UTC Normal Pod report-span-kbzjj AddedInterface Add eth0 [10.131.0.66/23] from ovn-kubernetes multus logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:25 +0000 UTC Normal Pod report-span-kbzjj.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:25 +0000 UTC Normal Pod report-span-kbzjj.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 356ms (356ms including waiting) kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:25 +0000 UTC Normal Pod report-span-kbzjj.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:25 +0000 UTC Normal Pod report-span-kbzjj.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 16:02:37 | examples-service-types | 2024-06-17 16:02:36 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 16:02:37 | examples-service-types | Deleting namespace: kuttl-test-thankful-akita === CONT kuttl/harness/examples-openshift-with-htpasswd logger.go:42: 16:03:04 | examples-openshift-with-htpasswd | Ignoring 00-install.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 16:03:04 | examples-openshift-with-htpasswd | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 16:03:04 | examples-openshift-with-htpasswd | Creating namespace: kuttl-test-great-oarfish logger.go:42: 16:03:05 | examples-openshift-with-htpasswd/0-install | starting test step 0-install logger.go:42: 16:03:05 | examples-openshift-with-htpasswd/0-install | Secret:kuttl-test-great-oarfish/htpasswd created logger.go:42: 16:03:05 | examples-openshift-with-htpasswd/0-install | test step completed 0-install logger.go:42: 16:03:05 | examples-openshift-with-htpasswd/1-install | starting test step 1-install logger.go:42: 16:03:05 | examples-openshift-with-htpasswd/1-install | Jaeger:kuttl-test-great-oarfish/with-htpasswd created logger.go:42: 16:03:12 | examples-openshift-with-htpasswd/1-install | test step completed 1-install logger.go:42: 16:03:12 | examples-openshift-with-htpasswd/2-check-unsecured | starting test step 2-check-unsecured logger.go:42: 16:03:12 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [./ensure-ingress-host.sh] logger.go:42: 16:03:12 | examples-openshift-with-htpasswd/2-check-unsecured | Checking the Ingress host value was populated logger.go:42: 16:03:12 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 0 logger.go:42: 16:03:12 | examples-openshift-with-htpasswd/2-check-unsecured | Hostname is with-htpasswd-kuttl-test-great-oarfish.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com logger.go:42: 16:03:12 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 16:03:12 | examples-openshift-with-htpasswd/2-check-unsecured | Checking an expected HTTP response logger.go:42: 16:03:12 | examples-openshift-with-htpasswd/2-check-unsecured | Running in OpenShift logger.go:42: 16:03:12 | examples-openshift-with-htpasswd/2-check-unsecured | Not using any secret logger.go:42: 16:03:12 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 1/30 the https://with-htpasswd-kuttl-test-great-oarfish.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 16:03:12 | examples-openshift-with-htpasswd/2-check-unsecured | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 16:03:12 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 2/30 the https://with-htpasswd-kuttl-test-great-oarfish.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 16:03:12 | examples-openshift-with-htpasswd/2-check-unsecured | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 16:03:22 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 3/30 the https://with-htpasswd-kuttl-test-great-oarfish.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 16:03:22 | examples-openshift-with-htpasswd/2-check-unsecured | curl response asserted properly logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/2-check-unsecured | test step completed 2-check-unsecured logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/3-check-unauthorized | starting test step 3-check-unauthorized logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [./ensure-ingress-host.sh] logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking the Ingress host value was populated logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 0 logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/3-check-unauthorized | Hostname is with-htpasswd-kuttl-test-great-oarfish.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [sh -c JAEGER_USERNAME=wronguser JAEGER_PASSWORD=wrongpassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking an expected HTTP response logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/3-check-unauthorized | Running in OpenShift logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/3-check-unauthorized | Using Jaeger basic authentication logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 1/30 the https://with-htpasswd-kuttl-test-great-oarfish.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/3-check-unauthorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 2/30 the https://with-htpasswd-kuttl-test-great-oarfish.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/3-check-unauthorized | curl response asserted properly logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/3-check-unauthorized | test step completed 3-check-unauthorized logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/4-check-authorized | starting test step 4-check-authorized logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/4-check-authorized | running command: [./ensure-ingress-host.sh] logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/4-check-authorized | Checking the Ingress host value was populated logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/4-check-authorized | Try number 0 logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/4-check-authorized | Hostname is with-htpasswd-kuttl-test-great-oarfish.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/4-check-authorized | running command: [sh -c JAEGER_USERNAME=awesomeuser JAEGER_PASSWORD=awesomepassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE with-htpasswd] logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/4-check-authorized | Checking an expected HTTP response logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/4-check-authorized | Running in OpenShift logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/4-check-authorized | Using Jaeger basic authentication logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/4-check-authorized | Try number 1/30 the https://with-htpasswd-kuttl-test-great-oarfish.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/4-check-authorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/4-check-authorized | Try number 2/30 the https://with-htpasswd-kuttl-test-great-oarfish.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/4-check-authorized | curl response asserted properly logger.go:42: 16:03:23 | examples-openshift-with-htpasswd/4-check-authorized | test step completed 4-check-authorized logger.go:42: 16:03:23 | examples-openshift-with-htpasswd | examples-openshift-with-htpasswd events from ns kuttl-test-great-oarfish: logger.go:42: 16:03:23 | examples-openshift-with-htpasswd | 2024-06-17 16:03:08 +0000 UTC Normal Pod with-htpasswd-684c978c9-r6jn7 Binding Scheduled Successfully assigned kuttl-test-great-oarfish/with-htpasswd-684c978c9-r6jn7 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:03:23 | examples-openshift-with-htpasswd | 2024-06-17 16:03:08 +0000 UTC Normal ReplicaSet.apps with-htpasswd-684c978c9 SuccessfulCreate Created pod: with-htpasswd-684c978c9-r6jn7 replicaset-controller logger.go:42: 16:03:23 | examples-openshift-with-htpasswd | 2024-06-17 16:03:08 +0000 UTC Normal Deployment.apps with-htpasswd ScalingReplicaSet Scaled up replica set with-htpasswd-684c978c9 to 1 deployment-controller logger.go:42: 16:03:23 | examples-openshift-with-htpasswd | 2024-06-17 16:03:09 +0000 UTC Normal Pod with-htpasswd-684c978c9-r6jn7 AddedInterface Add eth0 [10.129.2.45/23] from ovn-kubernetes multus logger.go:42: 16:03:23 | examples-openshift-with-htpasswd | 2024-06-17 16:03:09 +0000 UTC Normal Pod with-htpasswd-684c978c9-r6jn7.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:03:23 | examples-openshift-with-htpasswd | 2024-06-17 16:03:09 +0000 UTC Normal Pod with-htpasswd-684c978c9-r6jn7.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:03:23 | examples-openshift-with-htpasswd | 2024-06-17 16:03:10 +0000 UTC Normal Pod with-htpasswd-684c978c9-r6jn7.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:03:23 | examples-openshift-with-htpasswd | 2024-06-17 16:03:10 +0000 UTC Normal Pod with-htpasswd-684c978c9-r6jn7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:03:23 | examples-openshift-with-htpasswd | 2024-06-17 16:03:10 +0000 UTC Normal Pod with-htpasswd-684c978c9-r6jn7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:03:23 | examples-openshift-with-htpasswd | 2024-06-17 16:03:10 +0000 UTC Normal Pod with-htpasswd-684c978c9-r6jn7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:03:23 | examples-openshift-with-htpasswd | Deleting namespace: kuttl-test-great-oarfish === CONT kuttl/harness/examples-all-in-one-with-options logger.go:42: 16:03:30 | examples-all-in-one-with-options | Creating namespace: kuttl-test-winning-crayfish logger.go:42: 16:03:30 | examples-all-in-one-with-options/0-install | starting test step 0-install logger.go:42: 16:03:30 | examples-all-in-one-with-options/0-install | Jaeger:kuttl-test-winning-crayfish/my-jaeger created logger.go:42: 16:03:36 | examples-all-in-one-with-options/0-install | test step completed 0-install logger.go:42: 16:03:36 | examples-all-in-one-with-options/1-smoke-test | starting test step 1-smoke-test logger.go:42: 16:03:36 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 16:03:37 | examples-all-in-one-with-options/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 16:03:44 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443/jaeger MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 16:03:44 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 16:03:45 | examples-all-in-one-with-options/1-smoke-test | job.batch/report-span created logger.go:42: 16:03:45 | examples-all-in-one-with-options/1-smoke-test | job.batch/check-span created logger.go:42: 16:03:57 | examples-all-in-one-with-options/1-smoke-test | test step completed 1-smoke-test logger.go:42: 16:03:57 | examples-all-in-one-with-options | examples-all-in-one-with-options events from ns kuttl-test-winning-crayfish: logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:33 +0000 UTC Normal Pod my-jaeger-dbdc66d6c-dvrd5 Binding Scheduled Successfully assigned kuttl-test-winning-crayfish/my-jaeger-dbdc66d6c-dvrd5 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:33 +0000 UTC Normal ReplicaSet.apps my-jaeger-dbdc66d6c SuccessfulCreate Created pod: my-jaeger-dbdc66d6c-dvrd5 replicaset-controller logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:33 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-dbdc66d6c to 1 deployment-controller logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:34 +0000 UTC Normal Pod my-jaeger-dbdc66d6c-dvrd5 AddedInterface Add eth0 [10.129.2.46/23] from ovn-kubernetes multus logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:34 +0000 UTC Normal Pod my-jaeger-dbdc66d6c-dvrd5.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:34 +0000 UTC Normal Pod my-jaeger-dbdc66d6c-dvrd5.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:34 +0000 UTC Normal Pod my-jaeger-dbdc66d6c-dvrd5.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:34 +0000 UTC Normal Pod my-jaeger-dbdc66d6c-dvrd5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:34 +0000 UTC Normal Pod my-jaeger-dbdc66d6c-dvrd5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:34 +0000 UTC Normal Pod my-jaeger-dbdc66d6c-dvrd5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:39 +0000 UTC Normal Pod my-jaeger-dbdc66d6c-dvrd5.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:39 +0000 UTC Normal Pod my-jaeger-dbdc66d6c-dvrd5.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:39 +0000 UTC Normal ReplicaSet.apps my-jaeger-dbdc66d6c SuccessfulDelete Deleted pod: my-jaeger-dbdc66d6c-dvrd5 replicaset-controller logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:39 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-dbdc66d6c to 0 from 1 deployment-controller logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:40 +0000 UTC Normal Pod my-jaeger-6fdc647f58-b7ht6 Binding Scheduled Successfully assigned kuttl-test-winning-crayfish/my-jaeger-6fdc647f58-b7ht6 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:40 +0000 UTC Normal Pod my-jaeger-6fdc647f58-b7ht6 AddedInterface Add eth0 [10.129.2.47/23] from ovn-kubernetes multus logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:40 +0000 UTC Normal ReplicaSet.apps my-jaeger-6fdc647f58 SuccessfulCreate Created pod: my-jaeger-6fdc647f58-b7ht6 replicaset-controller logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:40 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6fdc647f58 to 1 deployment-controller logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:41 +0000 UTC Normal Pod my-jaeger-6fdc647f58-b7ht6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:41 +0000 UTC Normal Pod my-jaeger-6fdc647f58-b7ht6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:41 +0000 UTC Normal Pod my-jaeger-6fdc647f58-b7ht6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:41 +0000 UTC Normal Pod my-jaeger-6fdc647f58-b7ht6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:41 +0000 UTC Normal Pod my-jaeger-6fdc647f58-b7ht6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:41 +0000 UTC Normal Pod my-jaeger-6fdc647f58-b7ht6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:45 +0000 UTC Normal Pod check-span-59sjq Binding Scheduled Successfully assigned kuttl-test-winning-crayfish/check-span-59sjq to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:45 +0000 UTC Normal Pod check-span-59sjq AddedInterface Add eth0 [10.128.2.41/23] from ovn-kubernetes multus logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:45 +0000 UTC Normal Pod check-span-59sjq.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:45 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-59sjq job-controller logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:45 +0000 UTC Normal Pod report-span-z8cfg Binding Scheduled Successfully assigned kuttl-test-winning-crayfish/report-span-z8cfg to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:45 +0000 UTC Normal Pod report-span-z8cfg AddedInterface Add eth0 [10.131.0.67/23] from ovn-kubernetes multus logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:45 +0000 UTC Normal Pod report-span-z8cfg.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:45 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-z8cfg job-controller logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:46 +0000 UTC Normal Pod check-span-59sjq.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 380ms (380ms including waiting) kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:46 +0000 UTC Normal Pod check-span-59sjq.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:46 +0000 UTC Normal Pod check-span-59sjq.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:46 +0000 UTC Normal Pod report-span-z8cfg.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 339ms (339ms including waiting) kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:46 +0000 UTC Normal Pod report-span-z8cfg.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:46 +0000 UTC Normal Pod report-span-z8cfg.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 16:03:57 | examples-all-in-one-with-options | 2024-06-17 16:03:57 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 16:03:57 | examples-all-in-one-with-options | Deleting namespace: kuttl-test-winning-crayfish === CONT kuttl/harness/examples-agent-with-priority-class logger.go:42: 16:04:09 | examples-agent-with-priority-class | Creating namespace: kuttl-test-natural-moccasin logger.go:42: 16:04:09 | examples-agent-with-priority-class/0-install | starting test step 0-install logger.go:42: 16:04:10 | examples-agent-with-priority-class/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 16:04:10 | examples-agent-with-priority-class/0-install | ServiceAccount:kuttl-test-natural-moccasin/jaeger-agent-daemonset created logger.go:42: 16:04:10 | examples-agent-with-priority-class/0-install | test step completed 0-install logger.go:42: 16:04:10 | examples-agent-with-priority-class/1-add-policy | starting test step 1-add-policy logger.go:42: 16:04:10 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 16:04:10 | examples-agent-with-priority-class/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 16:04:10 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 16:04:15 | examples-agent-with-priority-class/1-add-policy | test step completed 1-add-policy logger.go:42: 16:04:15 | examples-agent-with-priority-class/2-install | starting test step 2-install logger.go:42: 16:04:15 | examples-agent-with-priority-class/2-install | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 16:04:23 | examples-agent-with-priority-class/2-install | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 16:04:23 | examples-agent-with-priority-class/2-install | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 16:04:24 | examples-agent-with-priority-class/2-install | job.batch/report-span created logger.go:42: 16:04:24 | examples-agent-with-priority-class/2-install | job.batch/check-span created logger.go:42: 16:04:24 | examples-agent-with-priority-class/2-install | PriorityClass:/high-priority created logger.go:42: 16:04:24 | examples-agent-with-priority-class/2-install | Jaeger:kuttl-test-natural-moccasin/agent-as-daemonset updated logger.go:42: 16:04:36 | examples-agent-with-priority-class/2-install | test step completed 2-install logger.go:42: 16:04:36 | examples-agent-with-priority-class | examples-agent-with-priority-class events from ns kuttl-test-natural-moccasin: logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:20 +0000 UTC Normal Pod agent-as-daemonset-84768d7d8-zzhgl Binding Scheduled Successfully assigned kuttl-test-natural-moccasin/agent-as-daemonset-84768d7d8-zzhgl to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:20 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-84768d7d8 SuccessfulCreate Created pod: agent-as-daemonset-84768d7d8-zzhgl replicaset-controller logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:20 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-84768d7d8 to 1 deployment-controller logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:21 +0000 UTC Normal Pod agent-as-daemonset-84768d7d8-zzhgl AddedInterface Add eth0 [10.129.2.48/23] from ovn-kubernetes multus logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:21 +0000 UTC Normal Pod agent-as-daemonset-84768d7d8-zzhgl.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:21 +0000 UTC Normal Pod agent-as-daemonset-84768d7d8-zzhgl.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:21 +0000 UTC Normal Pod agent-as-daemonset-84768d7d8-zzhgl.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:21 +0000 UTC Normal Pod agent-as-daemonset-84768d7d8-zzhgl.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:21 +0000 UTC Normal Pod agent-as-daemonset-84768d7d8-zzhgl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:21 +0000 UTC Normal Pod agent-as-daemonset-84768d7d8-zzhgl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:24 +0000 UTC Normal Pod check-span-8hb2b Binding Scheduled Successfully assigned kuttl-test-natural-moccasin/check-span-8hb2b to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:24 +0000 UTC Normal Pod check-span-8hb2b AddedInterface Add eth0 [10.128.2.42/23] from ovn-kubernetes multus logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:24 +0000 UTC Normal Pod check-span-8hb2b.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:24 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-8hb2b job-controller logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:24 +0000 UTC Normal Pod report-span-dcdtd Binding Scheduled Successfully assigned kuttl-test-natural-moccasin/report-span-dcdtd to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:24 +0000 UTC Normal Pod report-span-dcdtd AddedInterface Add eth0 [10.131.0.68/23] from ovn-kubernetes multus logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:24 +0000 UTC Normal Pod report-span-dcdtd.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:24 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-dcdtd job-controller logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:25 +0000 UTC Normal Pod check-span-8hb2b.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 360ms (360ms including waiting) kubelet logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:25 +0000 UTC Normal Pod check-span-8hb2b.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:25 +0000 UTC Normal Pod check-span-8hb2b.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:25 +0000 UTC Normal Pod report-span-dcdtd.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 344ms (344ms including waiting) kubelet logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:25 +0000 UTC Normal Pod report-span-dcdtd.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:25 +0000 UTC Normal Pod report-span-dcdtd.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:27 +0000 UTC Warning DaemonSet.apps agent-as-daemonset-agent-daemonset FailedCreate Error creating: pods "agent-as-daemonset-agent-daemonset-" is forbidden: unable to validate against any security context constraint: [provider "anyuid": Forbidden: not usable by user or serviceaccount, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5775: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5778: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6831: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6832: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 14271: Host ports are not allowed to be used, provider "restricted": Forbidden: not usable by user or serviceaccount, provider "nonroot-v2": Forbidden: not usable by user or serviceaccount, provider "nonroot": Forbidden: not usable by user or serviceaccount, provider "hostmount-anyuid": Forbidden: not usable by user or serviceaccount, provider "elasticsearch-scc": Forbidden: not usable by user or serviceaccount, provider "machine-api-termination-handler": Forbidden: not usable by user or serviceaccount, provider "daemonset-with-hostport": Forbidden: not usable by user or serviceaccount, provider "hostnetwork-v2": Forbidden: not usable by user or serviceaccount, provider "hostnetwork": Forbidden: not usable by user or serviceaccount, provider "hostaccess": Forbidden: not usable by user or serviceaccount, provider "node-exporter": Forbidden: not usable by user or serviceaccount, provider "privileged": Forbidden: not usable by user or serviceaccount] daemonset-controller logger.go:42: 16:04:36 | examples-agent-with-priority-class | 2024-06-17 16:04:36 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 16:04:36 | examples-agent-with-priority-class | Deleting namespace: kuttl-test-natural-moccasin === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (813.67s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.32s) --- PASS: kuttl/harness/examples-simple-prod (66.90s) --- PASS: kuttl/harness/examples-with-sampling (320.11s) --- PASS: kuttl/harness/examples-with-cassandra (74.94s) --- PASS: kuttl/harness/examples-with-badger (35.33s) --- PASS: kuttl/harness/examples-simplest (40.32s) --- PASS: kuttl/harness/examples-simple-prod-with-volumes (71.49s) --- PASS: kuttl/harness/examples-business-application-injected-sidecar (40.49s) --- PASS: kuttl/harness/examples-service-types (57.79s) --- PASS: kuttl/harness/examples-openshift-with-htpasswd (25.12s) --- PASS: kuttl/harness/examples-all-in-one-with-options (39.84s) --- PASS: kuttl/harness/examples-agent-with-priority-class (34.79s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name examples --report --output /logs/artifacts/examples.xml ./artifacts/kuttl-report.xml time="2024-06-17T16:04:45Z" level=debug msg="Setting a new name for the test suites" time="2024-06-17T16:04:45Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-06-17T16:04:45Z" level=debug msg="normalizing test case names" time="2024-06-17T16:04:45Z" level=debug msg="examples/artifacts -> examples_artifacts" time="2024-06-17T16:04:45Z" level=debug msg="examples/examples-simple-prod -> examples_examples_simple_prod" time="2024-06-17T16:04:45Z" level=debug msg="examples/examples-with-sampling -> examples_examples_with_sampling" time="2024-06-17T16:04:45Z" level=debug msg="examples/examples-with-cassandra -> examples_examples_with_cassandra" time="2024-06-17T16:04:45Z" level=debug msg="examples/examples-with-badger -> examples_examples_with_badger" time="2024-06-17T16:04:45Z" level=debug msg="examples/examples-simplest -> examples_examples_simplest" time="2024-06-17T16:04:45Z" level=debug msg="examples/examples-simple-prod-with-volumes -> examples_examples_simple_prod_with_volumes" time="2024-06-17T16:04:45Z" level=debug msg="examples/examples-business-application-injected-sidecar -> examples_examples_business_application_injected_sidecar" time="2024-06-17T16:04:45Z" level=debug msg="examples/examples-service-types -> examples_examples_service_types" time="2024-06-17T16:04:45Z" level=debug msg="examples/examples-openshift-with-htpasswd -> examples_examples_openshift_with_htpasswd" time="2024-06-17T16:04:45Z" level=debug msg="examples/examples-all-in-one-with-options -> examples_examples_all_in_one_with_options" time="2024-06-17T16:04:45Z" level=debug msg="examples/examples-agent-with-priority-class -> examples_examples_agent_with_priority_class" +---------------------------------------------------------+--------+ | NAME | RESULT | +---------------------------------------------------------+--------+ | examples_artifacts | passed | | examples_examples_simple_prod | passed | | examples_examples_with_sampling | passed | | examples_examples_with_cassandra | passed | | examples_examples_with_badger | passed | | examples_examples_simplest | passed | | examples_examples_simple_prod_with_volumes | passed | | examples_examples_business_application_injected_sidecar | passed | | examples_examples_service_types | passed | | examples_examples_openshift_with_htpasswd | passed | | examples_examples_all_in_one_with_options | passed | | examples_examples_agent_with_priority_class | passed | +---------------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh generate false true + '[' 3 -ne 3 ']' + test_suite_name=generate + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/generate.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-generate make[2]: Entering directory '/tmp/jaeger-tests' test -s /tmp/jaeger-tests/bin/operator-sdk || curl -sLo /tmp/jaeger-tests/bin/operator-sdk https://github.com/operator-framework/operator-sdk/releases/download/v1.32.0/operator-sdk_`go env GOOS`_`go env GOARCH` ./hack/install/install-golangci-lint.sh Installing golangci-lint golangci-lint 1.55.2 is installed already ./hack/install/install-goimports.sh Installing goimports Try 0... go install golang.org/x/tools/cmd/goimports@v0.1.12 >>>> Formatting code... ./.ci/format.sh >>>> Building... ./hack/install/install-dependencies.sh Installing go dependencies Try 0... go mod download GOOS= GOARCH= CGO_ENABLED=0 GO111MODULE=on go build -ldflags "-X "github.com/jaegertracing/jaeger-operator/pkg/version".version="1.57.0" -X "github.com/jaegertracing/jaeger-operator/pkg/version".buildDate=2024-06-17T16:04:46Z -X "github.com/jaegertracing/jaeger-operator/pkg/version".defaultJaeger="1.57.0"" -o "bin/jaeger-operator" main.go JAEGER_VERSION="1.57.0" ./tests/e2e/generate/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-14-130320 True False 46m Cluster version is 4.16.0-0.nightly-2024-06-14-130320' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-14-130320 True False 46m Cluster version is 4.16.0-0.nightly-2024-06-14-130320' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/generate/render.sh ++ export SUITE_DIR=./tests/e2e/generate ++ SUITE_DIR=./tests/e2e/generate ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/generate ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test generate 'This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 2 -ne 2 ']' + test_name=generate + message='This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/generate/_build + '[' _build '!=' _build ']' + rm -rf generate + warning 'generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed\e[0m' WAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running generate E2E tests' Running generate E2E tests + cd tests/e2e/generate/_build + set +e + KUBECONFIG=/tmp/kubeconfig-37736524 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 16:05:02 | artifacts | Creating namespace: kuttl-test-knowing-mustang logger.go:42: 16:05:02 | artifacts | artifacts events from ns kuttl-test-knowing-mustang: logger.go:42: 16:05:02 | artifacts | Deleting namespace: kuttl-test-knowing-mustang === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.33s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.14s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name generate --report --output /logs/artifacts/generate.xml ./artifacts/kuttl-report.xml time="2024-06-17T16:05:08Z" level=debug msg="Setting a new name for the test suites" time="2024-06-17T16:05:08Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-06-17T16:05:08Z" level=debug msg="normalizing test case names" time="2024-06-17T16:05:08Z" level=debug msg="generate/artifacts -> generate_artifacts" +--------------------+--------+ | NAME | RESULT | +--------------------+--------+ | generate_artifacts | passed | +--------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh upgrade false true + '[' 3 -ne 3 ']' + test_suite_name=upgrade + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/upgrade.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-upgrade make[2]: Entering directory '/tmp/jaeger-tests' make docker JAEGER_VERSION=1.57.1 IMG="quay.io//jaeger-operator:next" make[3]: Entering directory '/tmp/jaeger-tests' [ ! -z "true" ] || docker build --build-arg=GOPROXY= --build-arg=VERSION="1.57.0" --build-arg=JAEGER_VERSION=1.57.1 --build-arg=TARGETARCH= --build-arg VERSION_DATE=2024-06-17T16:05:09Z --build-arg VERSION_PKG="github.com/jaegertracing/jaeger-operator/pkg/version" -t "quay.io//jaeger-operator:next" . make[3]: Leaving directory '/tmp/jaeger-tests' touch build-e2e-upgrade-image SKIP_ES_EXTERNAL=true IMG=quay.io//jaeger-operator:"1.57.0" JAEGER_OPERATOR_VERSION="1.57.0" JAEGER_VERSION="1.57.0" ./tests/e2e/upgrade/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-14-130320 True False 46m Cluster version is 4.16.0-0.nightly-2024-06-14-130320' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-14-130320 True False 46m Cluster version is 4.16.0-0.nightly-2024-06-14-130320' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/upgrade/render.sh ++ export SUITE_DIR=./tests/e2e/upgrade ++ SUITE_DIR=./tests/e2e/upgrade ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/upgrade ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + export JAEGER_NAME + '[' true = true ']' + skip_test upgrade 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade + warning 'upgrade: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade: Test not supported in OpenShift\e[0m' WAR: upgrade: Test not supported in OpenShift + '[' true = true ']' + skip_test upgrade-from-latest-release 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade-from-latest-release + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade-from-latest-release + warning 'upgrade-from-latest-release: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade-from-latest-release: Test not supported in OpenShift\e[0m' WAR: upgrade-from-latest-release: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running upgrade E2E tests' Running upgrade E2E tests + cd tests/e2e/upgrade/_build + set +e + KUBECONFIG=/tmp/kubeconfig-37736524 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 16:05:10 | artifacts | Creating namespace: kuttl-test-light-stork logger.go:42: 16:05:10 | artifacts | artifacts events from ns kuttl-test-light-stork: logger.go:42: 16:05:10 | artifacts | Deleting namespace: kuttl-test-light-stork === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (7.22s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (7.02s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name upgrade --report --output /logs/artifacts/upgrade.xml ./artifacts/kuttl-report.xml time="2024-06-17T16:05:17Z" level=debug msg="Setting a new name for the test suites" time="2024-06-17T16:05:17Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-06-17T16:05:17Z" level=debug msg="normalizing test case names" time="2024-06-17T16:05:17Z" level=debug msg="upgrade/artifacts -> upgrade_artifacts" +-------------------+--------+ | NAME | RESULT | +-------------------+--------+ | upgrade_artifacts | passed | +-------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh sidecar false true + '[' 3 -ne 3 ']' + test_suite_name=sidecar + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/sidecar.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-sidecar make[2]: Entering directory '/tmp/jaeger-tests' ./tests/e2e/sidecar/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-14-130320 True False 46m Cluster version is 4.16.0-0.nightly-2024-06-14-130320' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-14-130320 True False 46m Cluster version is 4.16.0-0.nightly-2024-06-14-130320' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/sidecar/render.sh ++ export SUITE_DIR=./tests/e2e/sidecar ++ SUITE_DIR=./tests/e2e/sidecar ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/sidecar ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + jaeger_service_name=order + start_test sidecar-deployment + '[' 1 -ne 1 ']' + test_name=sidecar-deployment + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-deployment' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-deployment\e[0m' Rendering files for test sidecar-deployment + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build + '[' _build '!=' _build ']' + mkdir -p sidecar-deployment + cd sidecar-deployment + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-namespace + '[' 1 -ne 1 ']' + test_name=sidecar-namespace + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-namespace' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-namespace\e[0m' Rendering files for test sidecar-namespace + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-deployment + '[' sidecar-deployment '!=' _build ']' + cd .. + mkdir -p sidecar-namespace + cd sidecar-namespace + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-skip-webhook + '[' 1 -ne 1 ']' + test_name=sidecar-skip-webhook + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-skip-webhook' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-skip-webhook\e[0m' Rendering files for test sidecar-skip-webhook + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-namespace + '[' sidecar-namespace '!=' _build ']' + cd .. + mkdir -p sidecar-skip-webhook + cd sidecar-skip-webhook + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running sidecar E2E tests' Running sidecar E2E tests + cd tests/e2e/sidecar/_build + set +e + KUBECONFIG=/tmp/kubeconfig-37736524 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/sidecar-deployment === PAUSE kuttl/harness/sidecar-deployment === RUN kuttl/harness/sidecar-namespace === PAUSE kuttl/harness/sidecar-namespace === RUN kuttl/harness/sidecar-skip-webhook === PAUSE kuttl/harness/sidecar-skip-webhook === CONT kuttl/harness/artifacts logger.go:42: 16:05:25 | artifacts | Creating namespace: kuttl-test-exact-grouse logger.go:42: 16:05:25 | artifacts | artifacts events from ns kuttl-test-exact-grouse: logger.go:42: 16:05:25 | artifacts | Deleting namespace: kuttl-test-exact-grouse === CONT kuttl/harness/sidecar-namespace logger.go:42: 16:05:32 | sidecar-namespace | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 16:05:32 | sidecar-namespace | Creating namespace: kuttl-test-assuring-gorilla logger.go:42: 16:05:32 | sidecar-namespace/0-install | starting test step 0-install logger.go:42: 16:05:32 | sidecar-namespace/0-install | Jaeger:kuttl-test-assuring-gorilla/agent-as-sidecar created logger.go:42: 16:05:37 | sidecar-namespace/0-install | test step completed 0-install logger.go:42: 16:05:37 | sidecar-namespace/1-install | starting test step 1-install logger.go:42: 16:05:37 | sidecar-namespace/1-install | Deployment:kuttl-test-assuring-gorilla/vertx-create-span-sidecar created logger.go:42: 16:05:38 | sidecar-namespace/1-install | test step completed 1-install logger.go:42: 16:05:38 | sidecar-namespace/2-enable-injection | starting test step 2-enable-injection logger.go:42: 16:05:38 | sidecar-namespace/2-enable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="true"] logger.go:42: 16:05:38 | sidecar-namespace/2-enable-injection | namespace/kuttl-test-assuring-gorilla annotated logger.go:42: 16:05:44 | sidecar-namespace/2-enable-injection | test step completed 2-enable-injection logger.go:42: 16:05:44 | sidecar-namespace/3-find-service | starting test step 3-find-service logger.go:42: 16:05:45 | sidecar-namespace/3-find-service | Job:kuttl-test-assuring-gorilla/00-find-service created logger.go:42: 16:05:57 | sidecar-namespace/3-find-service | test step completed 3-find-service logger.go:42: 16:05:57 | sidecar-namespace/4-other-instance | starting test step 4-other-instance logger.go:42: 16:05:57 | sidecar-namespace/4-other-instance | Jaeger:kuttl-test-assuring-gorilla/agent-as-sidecar2 created logger.go:42: 16:06:19 | sidecar-namespace/4-other-instance | test step completed 4-other-instance logger.go:42: 16:06:19 | sidecar-namespace/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 16:06:20 | sidecar-namespace/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 16:06:20 | sidecar-namespace/6-find-service | starting test step 6-find-service logger.go:42: 16:06:20 | sidecar-namespace/6-find-service | Job:kuttl-test-assuring-gorilla/01-find-service created logger.go:42: 16:06:40 | sidecar-namespace/6-find-service | test step completed 6-find-service logger.go:42: 16:06:40 | sidecar-namespace/7-disable-injection | starting test step 7-disable-injection logger.go:42: 16:06:40 | sidecar-namespace/7-disable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="false"] logger.go:42: 16:06:40 | sidecar-namespace/7-disable-injection | namespace/kuttl-test-assuring-gorilla annotated logger.go:42: 16:06:41 | sidecar-namespace/7-disable-injection | test step completed 7-disable-injection logger.go:42: 16:06:41 | sidecar-namespace | sidecar-namespace events from ns kuttl-test-assuring-gorilla: logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:35 +0000 UTC Normal Pod agent-as-sidecar-6c5f9f66bd-9wdlg Binding Scheduled Successfully assigned kuttl-test-assuring-gorilla/agent-as-sidecar-6c5f9f66bd-9wdlg to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:35 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-6c5f9f66bd SuccessfulCreate Created pod: agent-as-sidecar-6c5f9f66bd-9wdlg replicaset-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:35 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-6c5f9f66bd to 1 deployment-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:36 +0000 UTC Normal Pod agent-as-sidecar-6c5f9f66bd-9wdlg AddedInterface Add eth0 [10.129.2.49/23] from ovn-kubernetes multus logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:36 +0000 UTC Normal Pod agent-as-sidecar-6c5f9f66bd-9wdlg.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:36 +0000 UTC Normal Pod agent-as-sidecar-6c5f9f66bd-9wdlg.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:36 +0000 UTC Normal Pod agent-as-sidecar-6c5f9f66bd-9wdlg.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:37 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-fws4r Binding Scheduled Successfully assigned kuttl-test-assuring-gorilla/vertx-create-span-sidecar-84d458b68c-fws4r to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:37 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-fws4r replicaset-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:37 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:38 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-fws4r AddedInterface Add eth0 [10.131.0.69/23] from ovn-kubernetes multus logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:38 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-fws4r.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:38 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-fws4r.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:38 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-fws4r.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:38 +0000 UTC Normal Pod vertx-create-span-sidecar-85b9d487dc-rv2wl Binding Scheduled Successfully assigned kuttl-test-assuring-gorilla/vertx-create-span-sidecar-85b9d487dc-rv2wl to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:38 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-85b9d487dc SuccessfulCreate Created pod: vertx-create-span-sidecar-85b9d487dc-rv2wl replicaset-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:38 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-85b9d487dc to 1 deployment-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:39 +0000 UTC Normal Pod vertx-create-span-sidecar-85b9d487dc-rv2wl AddedInterface Add eth0 [10.129.2.50/23] from ovn-kubernetes multus logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:39 +0000 UTC Normal Pod vertx-create-span-sidecar-85b9d487dc-rv2wl.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:39 +0000 UTC Normal Pod vertx-create-span-sidecar-85b9d487dc-rv2wl.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:39 +0000 UTC Normal Pod vertx-create-span-sidecar-85b9d487dc-rv2wl.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:39 +0000 UTC Normal Pod vertx-create-span-sidecar-85b9d487dc-rv2wl.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:44 +0000 UTC Normal Pod vertx-create-span-sidecar-85b9d487dc-rv2wl.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" in 4.182s (4.182s including waiting) kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:44 +0000 UTC Normal Pod vertx-create-span-sidecar-85b9d487dc-rv2wl.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:44 +0000 UTC Normal Pod vertx-create-span-sidecar-85b9d487dc-rv2wl.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:45 +0000 UTC Normal Pod 00-find-service-4bmx2 Binding Scheduled Successfully assigned kuttl-test-assuring-gorilla/00-find-service-4bmx2 to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:45 +0000 UTC Normal Pod 00-find-service-4bmx2 AddedInterface Add eth0 [10.128.2.43/23] from ovn-kubernetes multus logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:45 +0000 UTC Normal Pod 00-find-service-4bmx2.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:45 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-4bmx2 job-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:46 +0000 UTC Normal Pod 00-find-service-4bmx2.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 348ms (348ms including waiting) kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:46 +0000 UTC Normal Pod 00-find-service-4bmx2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:46 +0000 UTC Normal Pod 00-find-service-4bmx2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:46 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-fws4r.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.69:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:46 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-fws4r.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.69:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:47 +0000 UTC Warning Pod vertx-create-span-sidecar-85b9d487dc-rv2wl.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.50:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:47 +0000 UTC Warning Pod vertx-create-span-sidecar-85b9d487dc-rv2wl.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.50:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:49 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-fws4r.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:49 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-fws4r.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.69:8080/": read tcp 10.131.0.2:59116->10.131.0.69:8080: read: connection reset by peer kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:49 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-fws4r.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.69:8080/": dial tcp 10.131.0.69:8080: connect: connection refused kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:49 +0000 UTC Normal Pod vertx-create-span-sidecar-85b9d487dc-rv2wl.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:49 +0000 UTC Warning Pod vertx-create-span-sidecar-85b9d487dc-rv2wl.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.50:8080/": read tcp 10.129.2.2:42876->10.129.2.50:8080: read: connection reset by peer kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:49 +0000 UTC Warning Pod vertx-create-span-sidecar-85b9d487dc-rv2wl.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.50:8080/": dial tcp 10.129.2.50:8080: connect: connection refused kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:05:56 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:00 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-fws4r.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.69:8080/": read tcp 10.131.0.2:49616->10.131.0.69:8080: read: connection reset by peer kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:03 +0000 UTC Normal Pod agent-as-sidecar2-554c555d4-d8c6z Binding Scheduled Successfully assigned kuttl-test-assuring-gorilla/agent-as-sidecar2-554c555d4-d8c6z to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:03 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-554c555d4 SuccessfulCreate Created pod: agent-as-sidecar2-554c555d4-d8c6z replicaset-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:03 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-554c555d4 to 1 deployment-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:04 +0000 UTC Normal Pod agent-as-sidecar2-554c555d4-d8c6z AddedInterface Add eth0 [10.128.2.44/23] from ovn-kubernetes multus logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:04 +0000 UTC Normal Pod agent-as-sidecar2-554c555d4-d8c6z.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:18 +0000 UTC Normal Pod agent-as-sidecar2-554c555d4-d8c6z.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" in 14.125s (14.125s including waiting) kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:18 +0000 UTC Normal Pod agent-as-sidecar2-554c555d4-d8c6z.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:18 +0000 UTC Normal Pod agent-as-sidecar2-554c555d4-d8c6z.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:20 +0000 UTC Normal Pod 01-find-service-5ng2l Binding Scheduled Successfully assigned kuttl-test-assuring-gorilla/01-find-service-5ng2l to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:20 +0000 UTC Normal Pod 01-find-service-5ng2l AddedInterface Add eth0 [10.131.0.70/23] from ovn-kubernetes multus logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:20 +0000 UTC Normal Pod 01-find-service-5ng2l.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:20 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-5ng2l job-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:20 +0000 UTC Normal Pod agent-as-sidecar-6c5f9f66bd-9wdlg.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:20 +0000 UTC Normal Pod vertx-create-span-sidecar-6f79dc6b5f-8699f Binding Scheduled Successfully assigned kuttl-test-assuring-gorilla/vertx-create-span-sidecar-6f79dc6b5f-8699f to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:20 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6f79dc6b5f SuccessfulCreate Created pod: vertx-create-span-sidecar-6f79dc6b5f-8699f replicaset-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:20 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-fws4r replicaset-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:20 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:20 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6f79dc6b5f to 1 from 0 deployment-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:21 +0000 UTC Normal Pod 01-find-service-5ng2l.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 381ms (381ms including waiting) kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:21 +0000 UTC Normal Pod 01-find-service-5ng2l.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:21 +0000 UTC Normal Pod 01-find-service-5ng2l.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:21 +0000 UTC Normal Pod vertx-create-span-sidecar-6f79dc6b5f-8699f AddedInterface Add eth0 [10.128.2.45/23] from ovn-kubernetes multus logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:21 +0000 UTC Normal Pod vertx-create-span-sidecar-6f79dc6b5f-8699f.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:25 +0000 UTC Normal Pod vertx-create-span-sidecar-6f79dc6b5f-8699f.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.902s (3.902s including waiting) kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:25 +0000 UTC Normal Pod vertx-create-span-sidecar-6f79dc6b5f-8699f.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:25 +0000 UTC Normal Pod vertx-create-span-sidecar-6f79dc6b5f-8699f.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:25 +0000 UTC Normal Pod vertx-create-span-sidecar-6f79dc6b5f-8699f.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:25 +0000 UTC Normal Pod vertx-create-span-sidecar-6f79dc6b5f-8699f.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:25 +0000 UTC Normal Pod vertx-create-span-sidecar-6f79dc6b5f-8699f.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:33 +0000 UTC Warning Pod vertx-create-span-sidecar-6f79dc6b5f-8699f.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.45:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:33 +0000 UTC Warning Pod vertx-create-span-sidecar-6f79dc6b5f-8699f.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.45:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:35 +0000 UTC Normal Pod vertx-create-span-sidecar-6f79dc6b5f-8699f.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:35 +0000 UTC Warning Pod vertx-create-span-sidecar-6f79dc6b5f-8699f.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.45:8080/": read tcp 10.128.2.2:44786->10.128.2.45:8080: read: connection reset by peer kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:35 +0000 UTC Warning Pod vertx-create-span-sidecar-6f79dc6b5f-8699f.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.45:8080/": dial tcp 10.128.2.45:8080: connect: connection refused kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:35 +0000 UTC Normal Pod vertx-create-span-sidecar-6f79dc6b5f-8699f.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:39 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:40 +0000 UTC Normal Pod vertx-create-span-sidecar-854779f7fc-4nzwg Binding Scheduled Successfully assigned kuttl-test-assuring-gorilla/vertx-create-span-sidecar-854779f7fc-4nzwg to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:40 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-854779f7fc SuccessfulCreate Created pod: vertx-create-span-sidecar-854779f7fc-4nzwg replicaset-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:40 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-85b9d487dc SuccessfulDelete Deleted pod: vertx-create-span-sidecar-85b9d487dc-rv2wl replicaset-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:40 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-85b9d487dc to 0 from 1 deployment-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:40 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-854779f7fc to 1 from 0 deployment-controller logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:41 +0000 UTC Normal Pod vertx-create-span-sidecar-854779f7fc-4nzwg AddedInterface Add eth0 [10.131.0.71/23] from ovn-kubernetes multus logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:41 +0000 UTC Normal Pod vertx-create-span-sidecar-854779f7fc-4nzwg.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:41 +0000 UTC Normal Pod vertx-create-span-sidecar-854779f7fc-4nzwg.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 16:06:41 | sidecar-namespace | 2024-06-17 16:06:41 +0000 UTC Normal Pod vertx-create-span-sidecar-854779f7fc-4nzwg.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 16:06:42 | sidecar-namespace | Deleting namespace: kuttl-test-assuring-gorilla === CONT kuttl/harness/sidecar-skip-webhook logger.go:42: 16:06:48 | sidecar-skip-webhook | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 16:06:48 | sidecar-skip-webhook | Creating namespace: kuttl-test-informed-sawfish logger.go:42: 16:06:49 | sidecar-skip-webhook/0-install | starting test step 0-install logger.go:42: 16:06:49 | sidecar-skip-webhook/0-install | Jaeger:kuttl-test-informed-sawfish/agent-as-sidecar created logger.go:42: 16:06:55 | sidecar-skip-webhook/0-install | test step completed 0-install logger.go:42: 16:06:55 | sidecar-skip-webhook/1-install | starting test step 1-install logger.go:42: 16:06:55 | sidecar-skip-webhook/1-install | Deployment:kuttl-test-informed-sawfish/vertx-create-span-sidecar created logger.go:42: 16:06:57 | sidecar-skip-webhook/1-install | test step completed 1-install logger.go:42: 16:06:57 | sidecar-skip-webhook/2-add-anotation-and-label | starting test step 2-add-anotation-and-label logger.go:42: 16:06:57 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name=jaeger-operator --namespace kuttl-test-informed-sawfish] logger.go:42: 16:06:57 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar labeled logger.go:42: 16:06:57 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-informed-sawfish] logger.go:42: 16:06:57 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 16:06:57 | sidecar-skip-webhook/2-add-anotation-and-label | test step completed 2-add-anotation-and-label logger.go:42: 16:06:57 | sidecar-skip-webhook/3-remove-label | starting test step 3-remove-label logger.go:42: 16:06:57 | sidecar-skip-webhook/3-remove-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name- --namespace kuttl-test-informed-sawfish] logger.go:42: 16:06:58 | sidecar-skip-webhook/3-remove-label | deployment.apps/vertx-create-span-sidecar unlabeled logger.go:42: 16:07:00 | sidecar-skip-webhook/3-remove-label | test step completed 3-remove-label logger.go:42: 16:07:00 | sidecar-skip-webhook | sidecar-skip-webhook events from ns kuttl-test-informed-sawfish: logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:52 +0000 UTC Normal Pod agent-as-sidecar-7697666cf7-whjl6 Binding Scheduled Successfully assigned kuttl-test-informed-sawfish/agent-as-sidecar-7697666cf7-whjl6 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:52 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-7697666cf7 SuccessfulCreate Created pod: agent-as-sidecar-7697666cf7-whjl6 replicaset-controller logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:52 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-7697666cf7 to 1 deployment-controller logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:53 +0000 UTC Normal Pod agent-as-sidecar-7697666cf7-whjl6 AddedInterface Add eth0 [10.129.2.51/23] from ovn-kubernetes multus logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:53 +0000 UTC Normal Pod agent-as-sidecar-7697666cf7-whjl6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:53 +0000 UTC Normal Pod agent-as-sidecar-7697666cf7-whjl6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:53 +0000 UTC Normal Pod agent-as-sidecar-7697666cf7-whjl6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:55 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-v46m4 Binding Scheduled Successfully assigned kuttl-test-informed-sawfish/vertx-create-span-sidecar-84d458b68c-v46m4 to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:55 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-v46m4 replicaset-controller logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:55 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:56 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-v46m4 AddedInterface Add eth0 [10.131.0.72/23] from ovn-kubernetes multus logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:56 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-v46m4.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:56 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-v46m4.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:56 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-v46m4.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:57 +0000 UTC Normal Pod vertx-create-span-sidecar-768dd67654-kzqnq Binding Scheduled Successfully assigned kuttl-test-informed-sawfish/vertx-create-span-sidecar-768dd67654-kzqnq to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:57 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-768dd67654 SuccessfulCreate Created pod: vertx-create-span-sidecar-768dd67654-kzqnq replicaset-controller logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:57 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-768dd67654 to 1 deployment-controller logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:58 +0000 UTC Normal Pod vertx-create-span-sidecar-768dd67654-kzqnq AddedInterface Add eth0 [10.128.2.46/23] from ovn-kubernetes multus logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:58 +0000 UTC Normal Pod vertx-create-span-sidecar-768dd67654-kzqnq.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:58 +0000 UTC Normal Pod vertx-create-span-sidecar-768dd67654-kzqnq.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:58 +0000 UTC Normal Pod vertx-create-span-sidecar-768dd67654-kzqnq.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:58 +0000 UTC Normal Pod vertx-create-span-sidecar-768dd67654-kzqnq.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:58 +0000 UTC Normal Pod vertx-create-span-sidecar-768dd67654-kzqnq.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:07:00 | sidecar-skip-webhook | 2024-06-17 16:06:58 +0000 UTC Normal Pod vertx-create-span-sidecar-768dd67654-kzqnq.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:07:00 | sidecar-skip-webhook | Deleting namespace: kuttl-test-informed-sawfish === CONT kuttl/harness/sidecar-deployment logger.go:42: 16:07:06 | sidecar-deployment | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 16:07:06 | sidecar-deployment | Creating namespace: kuttl-test-holy-macaque logger.go:42: 16:07:06 | sidecar-deployment/0-install | starting test step 0-install logger.go:42: 16:07:06 | sidecar-deployment/0-install | Jaeger:kuttl-test-holy-macaque/agent-as-sidecar created logger.go:42: 16:07:12 | sidecar-deployment/0-install | test step completed 0-install logger.go:42: 16:07:12 | sidecar-deployment/1-install | starting test step 1-install logger.go:42: 16:07:12 | sidecar-deployment/1-install | Deployment:kuttl-test-holy-macaque/vertx-create-span-sidecar created logger.go:42: 16:07:15 | sidecar-deployment/1-install | test step completed 1-install logger.go:42: 16:07:15 | sidecar-deployment/2-enable-injection | starting test step 2-enable-injection logger.go:42: 16:07:15 | sidecar-deployment/2-enable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-holy-macaque] logger.go:42: 16:07:15 | sidecar-deployment/2-enable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 16:07:17 | sidecar-deployment/2-enable-injection | test step completed 2-enable-injection logger.go:42: 16:07:17 | sidecar-deployment/3-find-service | starting test step 3-find-service logger.go:42: 16:07:17 | sidecar-deployment/3-find-service | Job:kuttl-test-holy-macaque/00-find-service created logger.go:42: 16:07:29 | sidecar-deployment/3-find-service | test step completed 3-find-service logger.go:42: 16:07:29 | sidecar-deployment/4-other-instance | starting test step 4-other-instance logger.go:42: 16:07:29 | sidecar-deployment/4-other-instance | Jaeger:kuttl-test-holy-macaque/agent-as-sidecar2 created logger.go:42: 16:07:35 | sidecar-deployment/4-other-instance | test step completed 4-other-instance logger.go:42: 16:07:35 | sidecar-deployment/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 16:07:36 | sidecar-deployment/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 16:07:36 | sidecar-deployment/6-find-service | starting test step 6-find-service logger.go:42: 16:07:36 | sidecar-deployment/6-find-service | Job:kuttl-test-holy-macaque/01-find-service created logger.go:42: 16:07:56 | sidecar-deployment/6-find-service | test step completed 6-find-service logger.go:42: 16:07:56 | sidecar-deployment/7-disable-injection | starting test step 7-disable-injection logger.go:42: 16:07:56 | sidecar-deployment/7-disable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=false --namespace kuttl-test-holy-macaque] logger.go:42: 16:07:56 | sidecar-deployment/7-disable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 16:07:58 | sidecar-deployment/7-disable-injection | test step completed 7-disable-injection logger.go:42: 16:07:58 | sidecar-deployment | sidecar-deployment events from ns kuttl-test-holy-macaque: logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:10 +0000 UTC Normal Pod agent-as-sidecar-5ccbc69fbc-vt58d Binding Scheduled Successfully assigned kuttl-test-holy-macaque/agent-as-sidecar-5ccbc69fbc-vt58d to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:10 +0000 UTC Normal Pod agent-as-sidecar-5ccbc69fbc-vt58d AddedInterface Add eth0 [10.129.2.52/23] from ovn-kubernetes multus logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:10 +0000 UTC Normal Pod agent-as-sidecar-5ccbc69fbc-vt58d.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:10 +0000 UTC Normal Pod agent-as-sidecar-5ccbc69fbc-vt58d.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:10 +0000 UTC Normal Pod agent-as-sidecar-5ccbc69fbc-vt58d.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:10 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-5ccbc69fbc SuccessfulCreate Created pod: agent-as-sidecar-5ccbc69fbc-vt58d replicaset-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:10 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-5ccbc69fbc to 1 deployment-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:12 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-8wdvk Binding Scheduled Successfully assigned kuttl-test-holy-macaque/vertx-create-span-sidecar-84d458b68c-8wdvk to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:12 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-8wdvk replicaset-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:12 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:13 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-8wdvk AddedInterface Add eth0 [10.131.0.73/23] from ovn-kubernetes multus logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:13 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-8wdvk.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:13 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-8wdvk.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:13 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-8wdvk.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:15 +0000 UTC Normal Pod vertx-create-span-sidecar-557f6876d4-4chr7 Binding Scheduled Successfully assigned kuttl-test-holy-macaque/vertx-create-span-sidecar-557f6876d4-4chr7 to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:15 +0000 UTC Normal Pod vertx-create-span-sidecar-557f6876d4-4chr7 AddedInterface Add eth0 [10.128.2.47/23] from ovn-kubernetes multus logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:15 +0000 UTC Normal Pod vertx-create-span-sidecar-557f6876d4-4chr7.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:15 +0000 UTC Normal Pod vertx-create-span-sidecar-557f6876d4-4chr7.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:15 +0000 UTC Normal Pod vertx-create-span-sidecar-557f6876d4-4chr7.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:15 +0000 UTC Normal Pod vertx-create-span-sidecar-557f6876d4-4chr7.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:15 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-557f6876d4 SuccessfulCreate Created pod: vertx-create-span-sidecar-557f6876d4-4chr7 replicaset-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:15 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-557f6876d4 to 1 deployment-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:16 +0000 UTC Normal Pod vertx-create-span-sidecar-557f6876d4-4chr7.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:16 +0000 UTC Normal Pod vertx-create-span-sidecar-557f6876d4-4chr7.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:17 +0000 UTC Normal Pod 00-find-service-sm4p6 Binding Scheduled Successfully assigned kuttl-test-holy-macaque/00-find-service-sm4p6 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:17 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-sm4p6 job-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:18 +0000 UTC Normal Pod 00-find-service-sm4p6 AddedInterface Add eth0 [10.129.2.53/23] from ovn-kubernetes multus logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:18 +0000 UTC Normal Pod 00-find-service-sm4p6.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:18 +0000 UTC Normal Pod 00-find-service-sm4p6.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 358ms (358ms including waiting) kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:18 +0000 UTC Normal Pod 00-find-service-sm4p6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:18 +0000 UTC Normal Pod 00-find-service-sm4p6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:21 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-8wdvk.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.73:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:21 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-8wdvk.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.73:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:23 +0000 UTC Warning Pod vertx-create-span-sidecar-557f6876d4-4chr7.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.47:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:23 +0000 UTC Warning Pod vertx-create-span-sidecar-557f6876d4-4chr7.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.47:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:23 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-8wdvk.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:24 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-8wdvk.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.73:8080/": read tcp 10.131.0.2:48230->10.131.0.73:8080: read: connection reset by peer kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:24 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-8wdvk.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.73:8080/": dial tcp 10.131.0.73:8080: connect: connection refused kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:25 +0000 UTC Normal Pod vertx-create-span-sidecar-557f6876d4-4chr7.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:26 +0000 UTC Warning Pod vertx-create-span-sidecar-557f6876d4-4chr7.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.47:8080/": read tcp 10.128.2.2:44982->10.128.2.47:8080: read: connection reset by peer kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:26 +0000 UTC Warning Pod vertx-create-span-sidecar-557f6876d4-4chr7.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.47:8080/": dial tcp 10.128.2.47:8080: connect: connection refused kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:29 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:33 +0000 UTC Normal Pod agent-as-sidecar2-59b6566b4f-ll5k5 Binding Scheduled Successfully assigned kuttl-test-holy-macaque/agent-as-sidecar2-59b6566b4f-ll5k5 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:33 +0000 UTC Normal Pod agent-as-sidecar2-59b6566b4f-ll5k5 AddedInterface Add eth0 [10.129.2.54/23] from ovn-kubernetes multus logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:33 +0000 UTC Normal Pod agent-as-sidecar2-59b6566b4f-ll5k5.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:33 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-59b6566b4f SuccessfulCreate Created pod: agent-as-sidecar2-59b6566b4f-ll5k5 replicaset-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:33 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-59b6566b4f to 1 deployment-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:34 +0000 UTC Normal Pod agent-as-sidecar2-59b6566b4f-ll5k5.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:34 +0000 UTC Normal Pod agent-as-sidecar2-59b6566b4f-ll5k5.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:35 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-8wdvk.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.73:8080/": read tcp 10.131.0.2:42212->10.131.0.73:8080: read: connection reset by peer kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:36 +0000 UTC Normal Pod 01-find-service-8rgwf Binding Scheduled Successfully assigned kuttl-test-holy-macaque/01-find-service-8rgwf to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:36 +0000 UTC Normal Pod 01-find-service-8rgwf AddedInterface Add eth0 [10.131.0.74/23] from ovn-kubernetes multus logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:36 +0000 UTC Normal Pod 01-find-service-8rgwf.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:36 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-8rgwf job-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:36 +0000 UTC Normal Pod agent-as-sidecar-5ccbc69fbc-vt58d.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:37 +0000 UTC Normal Pod 01-find-service-8rgwf.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 435ms (435ms including waiting) kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:37 +0000 UTC Normal Pod 01-find-service-8rgwf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:37 +0000 UTC Normal Pod 01-find-service-8rgwf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:40 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-8wdvk replicaset-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:40 +0000 UTC Normal Pod vertx-create-span-sidecar-887cb8d8d-jdzdb Binding Scheduled Successfully assigned kuttl-test-holy-macaque/vertx-create-span-sidecar-887cb8d8d-jdzdb to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:40 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-887cb8d8d SuccessfulCreate Created pod: vertx-create-span-sidecar-887cb8d8d-jdzdb replicaset-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:40 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:40 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-887cb8d8d to 1 from 0 deployment-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:41 +0000 UTC Normal Pod vertx-create-span-sidecar-887cb8d8d-jdzdb AddedInterface Add eth0 [10.129.2.55/23] from ovn-kubernetes multus logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:41 +0000 UTC Normal Pod vertx-create-span-sidecar-887cb8d8d-jdzdb.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:41 +0000 UTC Normal Pod vertx-create-span-sidecar-887cb8d8d-jdzdb.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:41 +0000 UTC Normal Pod vertx-create-span-sidecar-887cb8d8d-jdzdb.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:41 +0000 UTC Normal Pod vertx-create-span-sidecar-887cb8d8d-jdzdb.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:41 +0000 UTC Normal Pod vertx-create-span-sidecar-887cb8d8d-jdzdb.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:41 +0000 UTC Normal Pod vertx-create-span-sidecar-887cb8d8d-jdzdb.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:49 +0000 UTC Warning Pod vertx-create-span-sidecar-887cb8d8d-jdzdb.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.55:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:49 +0000 UTC Warning Pod vertx-create-span-sidecar-887cb8d8d-jdzdb.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.55:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:51 +0000 UTC Normal Pod vertx-create-span-sidecar-887cb8d8d-jdzdb.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:51 +0000 UTC Warning Pod vertx-create-span-sidecar-887cb8d8d-jdzdb.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.55:8080/": read tcp 10.129.2.2:35486->10.129.2.55:8080: read: connection reset by peer kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:51 +0000 UTC Warning Pod vertx-create-span-sidecar-887cb8d8d-jdzdb.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.55:8080/": dial tcp 10.129.2.55:8080: connect: connection refused kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:55 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:56 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-557f6876d4 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-557f6876d4-4chr7 replicaset-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:56 +0000 UTC Normal Pod vertx-create-span-sidecar-57644bd5bb-wxv5l Binding Scheduled Successfully assigned kuttl-test-holy-macaque/vertx-create-span-sidecar-57644bd5bb-wxv5l to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:56 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-57644bd5bb SuccessfulCreate Created pod: vertx-create-span-sidecar-57644bd5bb-wxv5l replicaset-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:56 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-557f6876d4 to 0 from 1 deployment-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:56 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-57644bd5bb to 1 from 0 deployment-controller logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:57 +0000 UTC Normal Pod vertx-create-span-sidecar-57644bd5bb-wxv5l AddedInterface Add eth0 [10.131.0.75/23] from ovn-kubernetes multus logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:57 +0000 UTC Normal Pod vertx-create-span-sidecar-57644bd5bb-wxv5l.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:57 +0000 UTC Normal Pod vertx-create-span-sidecar-57644bd5bb-wxv5l.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 16:07:58 | sidecar-deployment | 2024-06-17 16:07:57 +0000 UTC Normal Pod vertx-create-span-sidecar-57644bd5bb-wxv5l.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 16:07:58 | sidecar-deployment | Deleting namespace: kuttl-test-holy-macaque === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (160.30s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.32s) --- PASS: kuttl/harness/sidecar-namespace (76.92s) --- PASS: kuttl/harness/sidecar-skip-webhook (17.52s) --- PASS: kuttl/harness/sidecar-deployment (59.37s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name sidecar --report --output /logs/artifacts/sidecar.xml ./artifacts/kuttl-report.xml time="2024-06-17T16:08:06Z" level=debug msg="Setting a new name for the test suites" time="2024-06-17T16:08:06Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-06-17T16:08:06Z" level=debug msg="normalizing test case names" time="2024-06-17T16:08:06Z" level=debug msg="sidecar/artifacts -> sidecar_artifacts" time="2024-06-17T16:08:06Z" level=debug msg="sidecar/sidecar-namespace -> sidecar_sidecar_namespace" time="2024-06-17T16:08:06Z" level=debug msg="sidecar/sidecar-skip-webhook -> sidecar_sidecar_skip_webhook" time="2024-06-17T16:08:06Z" level=debug msg="sidecar/sidecar-deployment -> sidecar_sidecar_deployment" +------------------------------+--------+ | NAME | RESULT | +------------------------------+--------+ | sidecar_artifacts | passed | | sidecar_sidecar_namespace | passed | | sidecar_sidecar_skip_webhook | passed | | sidecar_sidecar_deployment | passed | +------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh streaming false true + '[' 3 -ne 3 ']' + test_suite_name=streaming + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/streaming.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-streaming make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ SKIP_ES_EXTERNAL=true \ ./tests/e2e/streaming/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-14-130320 True False 49m Cluster version is 4.16.0-0.nightly-2024-06-14-130320' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-14-130320 True False 49m Cluster version is 4.16.0-0.nightly-2024-06-14-130320' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/streaming/render.sh ++ export SUITE_DIR=./tests/e2e/streaming ++ SUITE_DIR=./tests/e2e/streaming ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/streaming ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + '[' false = true ']' + start_test streaming-simple + '[' 1 -ne 1 ']' + test_name=streaming-simple + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-simple' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-simple\e[0m' Rendering files for test streaming-simple + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + mkdir -p streaming-simple + cd streaming-simple + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/streaming-jaeger-assert.yaml.template -o ./04-assert.yaml + render_smoke_test simple-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=simple-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + export JAEGER_NAME=simple-streaming + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-tls + '[' 1 -ne 1 ']' + test_name=streaming-with-tls + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-tls' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-tls\e[0m' Rendering files for test streaming-with-tls + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-simple + '[' streaming-simple '!=' _build ']' + cd .. + mkdir -p streaming-with-tls + cd streaming-with-tls + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + render_smoke_test tls-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=tls-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + export JAEGER_NAME=tls-streaming + JAEGER_NAME=tls-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-autoprovisioning-autoscale + '[' 1 -ne 1 ']' + test_name=streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-autoprovisioning-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-autoprovisioning-autoscale\e[0m' Rendering files for test streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-tls + '[' streaming-with-tls '!=' _build ']' + cd .. + mkdir -p streaming-with-autoprovisioning-autoscale + cd streaming-with-autoprovisioning-autoscale + '[' true = true ']' + rm ./00-install.yaml ./00-assert.yaml + render_install_elasticsearch upstream 01 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=01 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./01-assert.yaml + jaeger_name=auto-provisioned + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="20Mi"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="500m"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.autoscale=true ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.minReplicas=1 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.maxReplicas=2 ./02-install.yaml + render_assert_kafka true auto-provisioned 03 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provisioned + test_step=03 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./03-assert.yaml ++ expr 03 + 1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./04-assert.yaml ++ expr 03 + 2 + CLUSTER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./05-assert.yaml + version_lt 1.27 1.23 ++ echo 1.27 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.27 '!=' 1.27 + rm ./08-assert.yaml + skip_test streaming-with-tls 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-with-tls + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-autoprovisioning-autoscale + '[' streaming-with-autoprovisioning-autoscale '!=' _build ']' + cd .. + rm -rf streaming-with-tls + warning 'streaming-with-tls: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-with-tls: This test is flaky in Prow CI\e[0m' WAR: streaming-with-tls: This test is flaky in Prow CI + skip_test streaming-simple 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-simple + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + rm -rf streaming-simple + warning 'streaming-simple: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-simple: This test is flaky in Prow CI\e[0m' WAR: streaming-simple: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running streaming E2E tests' Running streaming E2E tests + cd tests/e2e/streaming/_build + set +e + KUBECONFIG=/tmp/kubeconfig-37736524 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 2 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/streaming-with-autoprovisioning-autoscale === PAUSE kuttl/harness/streaming-with-autoprovisioning-autoscale === CONT kuttl/harness/artifacts logger.go:42: 16:08:18 | artifacts | Creating namespace: kuttl-test-honest-horse logger.go:42: 16:08:18 | artifacts | artifacts events from ns kuttl-test-honest-horse: logger.go:42: 16:08:18 | artifacts | Deleting namespace: kuttl-test-honest-horse === CONT kuttl/harness/streaming-with-autoprovisioning-autoscale logger.go:42: 16:08:24 | streaming-with-autoprovisioning-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 16:08:24 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 16:08:24 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 16:08:24 | streaming-with-autoprovisioning-autoscale | Creating namespace: kuttl-test-charming-mosquito logger.go:42: 16:08:24 | streaming-with-autoprovisioning-autoscale/1-install | starting test step 1-install logger.go:42: 16:08:24 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc create sa deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 16:08:24 | streaming-with-autoprovisioning-autoscale/1-install | serviceaccount/deploy-elasticsearch created logger.go:42: 16:08:24 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc adm policy add-scc-to-user privileged -z deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 16:08:24 | streaming-with-autoprovisioning-autoscale/1-install | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:privileged added: "deploy-elasticsearch" logger.go:42: 16:08:24 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 6] logger.go:42: 16:08:30 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_0.yml -n $NAMESPACE] logger.go:42: 16:08:31 | streaming-with-autoprovisioning-autoscale/1-install | statefulset.apps/elasticsearch created logger.go:42: 16:08:31 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 3] logger.go:42: 16:08:34 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_1.yml -n $NAMESPACE] logger.go:42: 16:08:34 | streaming-with-autoprovisioning-autoscale/1-install | service/elasticsearch created logger.go:42: 16:08:52 | streaming-with-autoprovisioning-autoscale/1-install | test step completed 1-install logger.go:42: 16:08:52 | streaming-with-autoprovisioning-autoscale/2-install | starting test step 2-install logger.go:42: 16:08:52 | streaming-with-autoprovisioning-autoscale/2-install | Jaeger:kuttl-test-charming-mosquito/auto-provisioned created logger.go:42: 16:08:52 | streaming-with-autoprovisioning-autoscale/2-install | test step completed 2-install logger.go:42: 16:08:52 | streaming-with-autoprovisioning-autoscale/3- | starting test step 3- logger.go:42: 16:09:45 | streaming-with-autoprovisioning-autoscale/3- | test step completed 3- logger.go:42: 16:09:45 | streaming-with-autoprovisioning-autoscale/4- | starting test step 4- logger.go:42: 16:10:16 | streaming-with-autoprovisioning-autoscale/4- | test step completed 4- logger.go:42: 16:10:16 | streaming-with-autoprovisioning-autoscale/5- | starting test step 5- logger.go:42: 16:10:39 | streaming-with-autoprovisioning-autoscale/5- | test step completed 5- logger.go:42: 16:10:39 | streaming-with-autoprovisioning-autoscale/6- | starting test step 6- logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale/6- | test step completed 6- logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale/7- | starting test step 7- logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale/7- | test step completed 7- logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | streaming-with-autoprovisioning-autoscale events from ns kuttl-test-charming-mosquito: logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:08:31 +0000 UTC Normal Pod elasticsearch-0 Binding Scheduled Successfully assigned kuttl-test-charming-mosquito/elasticsearch-0 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:08:31 +0000 UTC Normal StatefulSet.apps elasticsearch SuccessfulCreate create Pod elasticsearch-0 in StatefulSet elasticsearch successful statefulset-controller logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:08:32 +0000 UTC Normal Pod elasticsearch-0 AddedInterface Add eth0 [10.129.2.56/23] from ovn-kubernetes multus logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:08:32 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulling Pulling image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:08:40 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulled Successfully pulled image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" in 7.781s (7.781s including waiting) kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:08:40 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:08:40 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:08:47 +0000 UTC Warning Pod elasticsearch-0.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Get "http://10.129.2.56:9200/": dial tcp 10.129.2.56:9200: connect: connection refused kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:08:59 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-zookeeper NoPods No matching pods found controllermanager logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:08:59 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:08:59 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:08:59 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-charming-mosquito/data-auto-provisioned-zookeeper-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-796d45b4b5-pgh25_b055d7b4-cf8b-471a-8f65-280d9834dae0 logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:04 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 Binding Scheduled Successfully assigned kuttl-test-charming-mosquito/auto-provisioned-zookeeper-0 to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:04 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ProvisioningSucceeded Successfully provisioned volume pvc-47e4027e-7aea-4a47-b35e-ea01f5964368 ebs.csi.aws.com_aws-ebs-csi-driver-controller-796d45b4b5-pgh25_b055d7b4-cf8b-471a-8f65-280d9834dae0 logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:07 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-47e4027e-7aea-4a47-b35e-ea01f5964368" attachdetach-controller logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:15 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 AddedInterface Add eth0 [10.131.0.76/23] from ovn-kubernetes multus logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:15 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:6d800bc0aa810993a93517c2464eb4575ee0d2579a1b0d84f3b8614550e32c94" kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:25 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:6d800bc0aa810993a93517c2464eb4575ee0d2579a1b0d84f3b8614550e32c94" in 10.646s (10.646s including waiting) kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:25 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Created Created container zookeeper kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:25 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Started Started container zookeeper kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:46 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:47 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-kafka NoPods No matching pods found controllermanager logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:47 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-charming-mosquito/data-0-auto-provisioned-kafka-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-796d45b4b5-pgh25_b055d7b4-cf8b-471a-8f65-280d9834dae0 logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:47 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:51 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ProvisioningSucceeded Successfully provisioned volume pvc-25989c80-b72b-4eab-a48b-3cc486f95a01 ebs.csi.aws.com_aws-ebs-csi-driver-controller-796d45b4b5-pgh25_b055d7b4-cf8b-471a-8f65-280d9834dae0 logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:52 +0000 UTC Normal Pod auto-provisioned-kafka-0 Binding Scheduled Successfully assigned kuttl-test-charming-mosquito/auto-provisioned-kafka-0 to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:54 +0000 UTC Normal Pod auto-provisioned-kafka-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-25989c80-b72b-4eab-a48b-3cc486f95a01" attachdetach-controller logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:56 +0000 UTC Normal Pod auto-provisioned-kafka-0 AddedInterface Add eth0 [10.131.0.77/23] from ovn-kubernetes multus logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:56 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulled Container image "registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:6d800bc0aa810993a93517c2464eb4575ee0d2579a1b0d84f3b8614550e32c94" already present on machine kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:56 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Created Created container kafka kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:09:56 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Started Started container kafka kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:18 +0000 UTC Normal Pod auto-provisioned-entity-operator-594845b677-dwtp8 Binding Scheduled Successfully assigned kuttl-test-charming-mosquito/auto-provisioned-entity-operator-594845b677-dwtp8 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:18 +0000 UTC Normal ReplicaSet.apps auto-provisioned-entity-operator-594845b677 SuccessfulCreate Created pod: auto-provisioned-entity-operator-594845b677-dwtp8 replicaset-controller logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:18 +0000 UTC Normal Deployment.apps auto-provisioned-entity-operator ScalingReplicaSet Scaled up replica set auto-provisioned-entity-operator-594845b677 to 1 deployment-controller logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:19 +0000 UTC Normal Pod auto-provisioned-entity-operator-594845b677-dwtp8 AddedInterface Add eth0 [10.129.2.57/23] from ovn-kubernetes multus logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:19 +0000 UTC Normal Pod auto-provisioned-entity-operator-594845b677-dwtp8.spec.containers{topic-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:25768564d6d2b15ac00c95e595878716001d49bf896599d31b4fa6e17b64e8c4" already present on machine kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:19 +0000 UTC Normal Pod auto-provisioned-entity-operator-594845b677-dwtp8.spec.containers{topic-operator} Created Created container topic-operator kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:19 +0000 UTC Normal Pod auto-provisioned-entity-operator-594845b677-dwtp8.spec.containers{topic-operator} Started Started container topic-operator kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:19 +0000 UTC Normal Pod auto-provisioned-entity-operator-594845b677-dwtp8.spec.containers{user-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:25768564d6d2b15ac00c95e595878716001d49bf896599d31b4fa6e17b64e8c4" already present on machine kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:19 +0000 UTC Normal Pod auto-provisioned-entity-operator-594845b677-dwtp8.spec.containers{user-operator} Created Created container user-operator kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:19 +0000 UTC Normal Pod auto-provisioned-entity-operator-594845b677-dwtp8.spec.containers{user-operator} Started Started container user-operator kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:41 +0000 UTC Normal Pod auto-provisioned-collector-5d4645484c-ft8z2 Binding Scheduled Successfully assigned kuttl-test-charming-mosquito/auto-provisioned-collector-5d4645484c-ft8z2 to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:41 +0000 UTC Normal Pod auto-provisioned-collector-5d4645484c-ft8z2 AddedInterface Add eth0 [10.128.2.48/23] from ovn-kubernetes multus logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:41 +0000 UTC Normal Pod auto-provisioned-collector-5d4645484c-ft8z2.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1" already present on machine kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:41 +0000 UTC Normal ReplicaSet.apps auto-provisioned-collector-5d4645484c SuccessfulCreate Created pod: auto-provisioned-collector-5d4645484c-ft8z2 replicaset-controller logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:41 +0000 UTC Normal Deployment.apps auto-provisioned-collector ScalingReplicaSet Scaled up replica set auto-provisioned-collector-5d4645484c to 1 deployment-controller logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:41 +0000 UTC Normal Pod auto-provisioned-ingester-854dc5b494-qwtwb Binding Scheduled Successfully assigned kuttl-test-charming-mosquito/auto-provisioned-ingester-854dc5b494-qwtwb to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:41 +0000 UTC Normal Pod auto-provisioned-ingester-854dc5b494-qwtwb AddedInterface Add eth0 [10.128.2.49/23] from ovn-kubernetes multus logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:41 +0000 UTC Normal Pod auto-provisioned-ingester-854dc5b494-qwtwb.spec.containers{jaeger-ingester} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:92551a79508e01d08c260f6430ad6fe6851fd0c1d7e30d9316cfeadcda960b06" kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:41 +0000 UTC Normal ReplicaSet.apps auto-provisioned-ingester-854dc5b494 SuccessfulCreate Created pod: auto-provisioned-ingester-854dc5b494-qwtwb replicaset-controller logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:41 +0000 UTC Normal Deployment.apps auto-provisioned-ingester ScalingReplicaSet Scaled up replica set auto-provisioned-ingester-854dc5b494 to 1 deployment-controller logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:41 +0000 UTC Normal Pod auto-provisioned-query-9dcd69ffb-bdss5 Binding Scheduled Successfully assigned kuttl-test-charming-mosquito/auto-provisioned-query-9dcd69ffb-bdss5 to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:41 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-9dcd69ffb SuccessfulCreate Created pod: auto-provisioned-query-9dcd69ffb-bdss5 replicaset-controller logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:41 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled up replica set auto-provisioned-query-9dcd69ffb to 1 deployment-controller logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:42 +0000 UTC Normal Pod auto-provisioned-collector-5d4645484c-ft8z2.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:42 +0000 UTC Normal Pod auto-provisioned-collector-5d4645484c-ft8z2.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:42 +0000 UTC Normal Pod auto-provisioned-query-9dcd69ffb-bdss5 AddedInterface Add eth0 [10.131.0.78/23] from ovn-kubernetes multus logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:42 +0000 UTC Normal Pod auto-provisioned-query-9dcd69ffb-bdss5.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:42 +0000 UTC Normal Pod auto-provisioned-query-9dcd69ffb-bdss5.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:42 +0000 UTC Normal Pod auto-provisioned-query-9dcd69ffb-bdss5.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:42 +0000 UTC Normal Pod auto-provisioned-query-9dcd69ffb-bdss5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:42 +0000 UTC Normal Pod auto-provisioned-query-9dcd69ffb-bdss5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:42 +0000 UTC Normal Pod auto-provisioned-query-9dcd69ffb-bdss5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:42 +0000 UTC Normal Pod auto-provisioned-query-9dcd69ffb-bdss5.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:42 +0000 UTC Normal Pod auto-provisioned-query-9dcd69ffb-bdss5.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:42 +0000 UTC Normal Pod auto-provisioned-query-9dcd69ffb-bdss5.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:50 +0000 UTC Normal Pod auto-provisioned-ingester-854dc5b494-qwtwb.spec.containers{jaeger-ingester} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:92551a79508e01d08c260f6430ad6fe6851fd0c1d7e30d9316cfeadcda960b06" in 8.262s (8.262s including waiting) kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:50 +0000 UTC Normal Pod auto-provisioned-ingester-854dc5b494-qwtwb.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | 2024-06-17 16:10:50 +0000 UTC Normal Pod auto-provisioned-ingester-854dc5b494-qwtwb.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 16:10:52 | streaming-with-autoprovisioning-autoscale | Deleting namespace: kuttl-test-charming-mosquito === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (189.66s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.13s) --- PASS: kuttl/harness/streaming-with-autoprovisioning-autoscale (183.34s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name streaming --report --output /logs/artifacts/streaming.xml ./artifacts/kuttl-report.xml time="2024-06-17T16:11:28Z" level=debug msg="Setting a new name for the test suites" time="2024-06-17T16:11:28Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-06-17T16:11:28Z" level=debug msg="normalizing test case names" time="2024-06-17T16:11:28Z" level=debug msg="streaming/artifacts -> streaming_artifacts" time="2024-06-17T16:11:28Z" level=debug msg="streaming/streaming-with-autoprovisioning-autoscale -> streaming_streaming_with_autoprovisioning_autoscale" +-----------------------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------------------+--------+ | streaming_artifacts | passed | | streaming_streaming_with_autoprovisioning_autoscale | passed | +-----------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh ui false true + '[' 3 -ne 3 ']' + test_suite_name=ui + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/ui.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-ui make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true ./tests/e2e/ui/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-14-130320 True False 52m Cluster version is 4.16.0-0.nightly-2024-06-14-130320' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-14-130320 True False 52m Cluster version is 4.16.0-0.nightly-2024-06-14-130320' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/ui/render.sh ++ export SUITE_DIR=./tests/e2e/ui ++ SUITE_DIR=./tests/e2e/ui ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/ui ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test allinone + '[' 1 -ne 1 ']' + test_name=allinone + echo =========================================================================== =========================================================================== + info 'Rendering files for test allinone' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test allinone\e[0m' Rendering files for test allinone + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build + '[' _build '!=' _build ']' + mkdir -p allinone + cd allinone + export GET_URL_COMMAND + export URL + export JAEGER_NAME=all-in-one-ui + JAEGER_NAME=all-in-one-ui + '[' true = true ']' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./01-curl.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./04-test-ui-config.yaml + start_test production + '[' 1 -ne 1 ']' + test_name=production + echo =========================================================================== =========================================================================== + info 'Rendering files for test production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test production\e[0m' Rendering files for test production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build/allinone + '[' allinone '!=' _build ']' + cd .. + mkdir -p production + cd production + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + [[ true = true ]] + [[ true = true ]] + render_install_jaeger production-ui production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + '[' true = true ']' + INSECURE=true + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-forbbiden-access.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-curl.yaml + INSECURE=true + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./05-check-disabled-security.yaml + ASSERT_PRESENT=false + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./06-check-NO-gaID.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./08-check-gaID.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running ui E2E tests' Running ui E2E tests + cd tests/e2e/ui/_build + set +e + KUBECONFIG=/tmp/kubeconfig-37736524 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 3 tests === RUN kuttl/harness === RUN kuttl/harness/allinone === PAUSE kuttl/harness/allinone === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/production === PAUSE kuttl/harness/production === CONT kuttl/harness/allinone logger.go:42: 16:11:34 | allinone | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 16:11:34 | allinone | Creating namespace: kuttl-test-crack-giraffe logger.go:42: 16:11:34 | allinone/0-install | starting test step 0-install logger.go:42: 16:11:34 | allinone/0-install | Jaeger:kuttl-test-crack-giraffe/all-in-one-ui created logger.go:42: 16:11:39 | allinone/0-install | test step completed 0-install logger.go:42: 16:11:39 | allinone/1-curl | starting test step 1-curl logger.go:42: 16:11:39 | allinone/1-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 16:11:39 | allinone/1-curl | Checking the Ingress host value was populated logger.go:42: 16:11:39 | allinone/1-curl | Try number 0 logger.go:42: 16:11:39 | allinone/1-curl | Hostname is all-in-one-ui-kuttl-test-crack-giraffe.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com logger.go:42: 16:11:39 | allinone/1-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE all-in-one-ui] logger.go:42: 16:11:39 | allinone/1-curl | Checking an expected HTTP response logger.go:42: 16:11:39 | allinone/1-curl | Running in OpenShift logger.go:42: 16:11:39 | allinone/1-curl | User not provided. Getting the token... logger.go:42: 16:11:40 | allinone/1-curl | Warning: resource jaegers/all-in-one-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 16:11:47 | allinone/1-curl | Try number 1/30 the https://all-in-one-ui-kuttl-test-crack-giraffe.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 16:11:47 | allinone/1-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 16:11:47 | allinone/1-curl | Try number 2/30 the https://all-in-one-ui-kuttl-test-crack-giraffe.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 16:11:47 | allinone/1-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 16:11:57 | allinone/1-curl | Try number 3/30 the https://all-in-one-ui-kuttl-test-crack-giraffe.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 16:11:57 | allinone/1-curl | curl response asserted properly logger.go:42: 16:11:57 | allinone/1-curl | test step completed 1-curl logger.go:42: 16:11:57 | allinone/2-delete | starting test step 2-delete logger.go:42: 16:11:57 | allinone/2-delete | Jaeger:kuttl-test-crack-giraffe/all-in-one-ui created logger.go:42: 16:11:57 | allinone/2-delete | test step completed 2-delete logger.go:42: 16:11:57 | allinone/3-install | starting test step 3-install logger.go:42: 16:11:57 | allinone/3-install | Jaeger:kuttl-test-crack-giraffe/all-in-one-ui updated logger.go:42: 16:11:57 | allinone/3-install | test step completed 3-install logger.go:42: 16:11:57 | allinone/4-test-ui-config | starting test step 4-test-ui-config logger.go:42: 16:11:57 | allinone/4-test-ui-config | running command: [./ensure-ingress-host.sh] logger.go:42: 16:11:57 | allinone/4-test-ui-config | Checking the Ingress host value was populated logger.go:42: 16:11:57 | allinone/4-test-ui-config | Try number 0 logger.go:42: 16:11:57 | allinone/4-test-ui-config | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 16:11:57 | allinone/4-test-ui-config | template was: logger.go:42: 16:11:57 | allinone/4-test-ui-config | {.items[0].status.ingress[0].host} logger.go:42: 16:11:57 | allinone/4-test-ui-config | object given to jsonpath engine was: logger.go:42: 16:11:57 | allinone/4-test-ui-config | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 16:11:57 | allinone/4-test-ui-config | logger.go:42: 16:11:57 | allinone/4-test-ui-config | logger.go:42: 16:12:07 | allinone/4-test-ui-config | Try number 1 logger.go:42: 16:12:07 | allinone/4-test-ui-config | Hostname is all-in-one-ui-kuttl-test-crack-giraffe.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com logger.go:42: 16:12:07 | allinone/4-test-ui-config | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 16:12:08 | allinone/4-test-ui-config | time="2024-06-17T16:12:08Z" level=info msg="Querying https://all-in-one-ui-kuttl-test-crack-giraffe.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search..." logger.go:42: 16:12:08 | allinone/4-test-ui-config | time="2024-06-17T16:12:08Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 16:12:08 | allinone/4-test-ui-config | time="2024-06-17T16:12:08Z" level=info msg="Polling to https://all-in-one-ui-kuttl-test-crack-giraffe.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search" logger.go:42: 16:12:08 | allinone/4-test-ui-config | time="2024-06-17T16:12:08Z" level=info msg="Doing request number 0" logger.go:42: 16:12:08 | allinone/4-test-ui-config | time="2024-06-17T16:12:08Z" level=info msg="Content found and asserted!" logger.go:42: 16:12:08 | allinone/4-test-ui-config | time="2024-06-17T16:12:08Z" level=info msg="Success!" logger.go:42: 16:12:08 | allinone/4-test-ui-config | test step completed 4-test-ui-config logger.go:42: 16:12:08 | allinone | allinone events from ns kuttl-test-crack-giraffe: logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:38 +0000 UTC Normal Pod all-in-one-ui-56df8bf5b-qjg67 Binding Scheduled Successfully assigned kuttl-test-crack-giraffe/all-in-one-ui-56df8bf5b-qjg67 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:38 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-56df8bf5b SuccessfulCreate Created pod: all-in-one-ui-56df8bf5b-qjg67 replicaset-controller logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:38 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-56df8bf5b to 1 deployment-controller logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:39 +0000 UTC Normal Pod all-in-one-ui-56df8bf5b-qjg67 AddedInterface Add eth0 [10.129.2.58/23] from ovn-kubernetes multus logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:39 +0000 UTC Normal Pod all-in-one-ui-56df8bf5b-qjg67.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:39 +0000 UTC Normal Pod all-in-one-ui-56df8bf5b-qjg67.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:39 +0000 UTC Normal Pod all-in-one-ui-56df8bf5b-qjg67.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:39 +0000 UTC Normal Pod all-in-one-ui-56df8bf5b-qjg67.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:39 +0000 UTC Normal Pod all-in-one-ui-56df8bf5b-qjg67.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:39 +0000 UTC Normal Pod all-in-one-ui-56df8bf5b-qjg67.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:43 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-56df8bf5b SuccessfulDelete Deleted pod: all-in-one-ui-56df8bf5b-qjg67 replicaset-controller logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:43 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-56df8bf5b to 0 from 1 deployment-controller logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:44 +0000 UTC Normal Pod all-in-one-ui-56df8bf5b-qjg67.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:44 +0000 UTC Normal Pod all-in-one-ui-56df8bf5b-qjg67.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:45 +0000 UTC Normal Pod all-in-one-ui-854fc7d478-vk7z8 Binding Scheduled Successfully assigned kuttl-test-crack-giraffe/all-in-one-ui-854fc7d478-vk7z8 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:45 +0000 UTC Normal Pod all-in-one-ui-854fc7d478-vk7z8 AddedInterface Add eth0 [10.129.2.59/23] from ovn-kubernetes multus logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:45 +0000 UTC Normal Pod all-in-one-ui-854fc7d478-vk7z8.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:45 +0000 UTC Normal Pod all-in-one-ui-854fc7d478-vk7z8.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:45 +0000 UTC Normal Pod all-in-one-ui-854fc7d478-vk7z8.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:45 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-854fc7d478 SuccessfulCreate Created pod: all-in-one-ui-854fc7d478-vk7z8 replicaset-controller logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:45 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-854fc7d478 to 1 deployment-controller logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:46 +0000 UTC Normal Pod all-in-one-ui-854fc7d478-vk7z8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:46 +0000 UTC Normal Pod all-in-one-ui-854fc7d478-vk7z8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:46 +0000 UTC Normal Pod all-in-one-ui-854fc7d478-vk7z8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:57 +0000 UTC Normal Pod all-in-one-ui-854fc7d478-vk7z8.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 16:12:08 | allinone | 2024-06-17 16:11:57 +0000 UTC Normal Pod all-in-one-ui-854fc7d478-vk7z8.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 16:12:08 | allinone | 2024-06-17 16:12:03 +0000 UTC Normal Pod all-in-one-ui-6ff7bb55f5-j2r4r Binding Scheduled Successfully assigned kuttl-test-crack-giraffe/all-in-one-ui-6ff7bb55f5-j2r4r to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:12:08 | allinone | 2024-06-17 16:12:03 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-6ff7bb55f5 SuccessfulCreate Created pod: all-in-one-ui-6ff7bb55f5-j2r4r replicaset-controller logger.go:42: 16:12:08 | allinone | 2024-06-17 16:12:03 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-6ff7bb55f5 to 1 deployment-controller logger.go:42: 16:12:08 | allinone | 2024-06-17 16:12:04 +0000 UTC Normal Pod all-in-one-ui-6ff7bb55f5-j2r4r AddedInterface Add eth0 [10.129.2.60/23] from ovn-kubernetes multus logger.go:42: 16:12:08 | allinone | 2024-06-17 16:12:04 +0000 UTC Normal Pod all-in-one-ui-6ff7bb55f5-j2r4r.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:12:08 | allinone | 2024-06-17 16:12:04 +0000 UTC Normal Pod all-in-one-ui-6ff7bb55f5-j2r4r.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:12:08 | allinone | 2024-06-17 16:12:04 +0000 UTC Normal Pod all-in-one-ui-6ff7bb55f5-j2r4r.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:12:08 | allinone | Deleting namespace: kuttl-test-crack-giraffe === CONT kuttl/harness/production logger.go:42: 16:12:16 | production | Ignoring add-tracking-id.yaml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 16:12:16 | production | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 16:12:16 | production | Creating namespace: kuttl-test-secure-poodle logger.go:42: 16:12:16 | production/1-install | starting test step 1-install logger.go:42: 16:12:16 | production/1-install | Jaeger:kuttl-test-secure-poodle/production-ui created logger.go:42: 16:12:52 | production/1-install | test step completed 1-install logger.go:42: 16:12:52 | production/2-check-forbbiden-access | starting test step 2-check-forbbiden-access logger.go:42: 16:12:52 | production/2-check-forbbiden-access | running command: [./ensure-ingress-host.sh] logger.go:42: 16:12:52 | production/2-check-forbbiden-access | Checking the Ingress host value was populated logger.go:42: 16:12:52 | production/2-check-forbbiden-access | Try number 0 logger.go:42: 16:12:53 | production/2-check-forbbiden-access | Hostname is production-ui-kuttl-test-secure-poodle.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com logger.go:42: 16:12:53 | production/2-check-forbbiden-access | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE production-ui] logger.go:42: 16:12:53 | production/2-check-forbbiden-access | Checking an expected HTTP response logger.go:42: 16:12:53 | production/2-check-forbbiden-access | Running in OpenShift logger.go:42: 16:12:53 | production/2-check-forbbiden-access | Not using any secret logger.go:42: 16:12:53 | production/2-check-forbbiden-access | Try number 1/30 the https://production-ui-kuttl-test-secure-poodle.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 16:12:53 | production/2-check-forbbiden-access | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 16:12:53 | production/2-check-forbbiden-access | Try number 2/30 the https://production-ui-kuttl-test-secure-poodle.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 16:12:53 | production/2-check-forbbiden-access | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 16:13:03 | production/2-check-forbbiden-access | Try number 3/30 the https://production-ui-kuttl-test-secure-poodle.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 16:13:03 | production/2-check-forbbiden-access | curl response asserted properly logger.go:42: 16:13:03 | production/2-check-forbbiden-access | test step completed 2-check-forbbiden-access logger.go:42: 16:13:03 | production/3-curl | starting test step 3-curl logger.go:42: 16:13:03 | production/3-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 16:13:03 | production/3-curl | Checking the Ingress host value was populated logger.go:42: 16:13:03 | production/3-curl | Try number 0 logger.go:42: 16:13:03 | production/3-curl | Hostname is production-ui-kuttl-test-secure-poodle.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com logger.go:42: 16:13:03 | production/3-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 16:13:03 | production/3-curl | Checking an expected HTTP response logger.go:42: 16:13:03 | production/3-curl | Running in OpenShift logger.go:42: 16:13:03 | production/3-curl | User not provided. Getting the token... logger.go:42: 16:13:05 | production/3-curl | Warning: resource jaegers/production-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 16:13:11 | production/3-curl | Try number 1/30 the https://production-ui-kuttl-test-secure-poodle.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 16:13:11 | production/3-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 16:13:11 | production/3-curl | Try number 2/30 the https://production-ui-kuttl-test-secure-poodle.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 16:13:11 | production/3-curl | curl response asserted properly logger.go:42: 16:13:11 | production/3-curl | test step completed 3-curl logger.go:42: 16:13:11 | production/4-install | starting test step 4-install logger.go:42: 16:13:11 | production/4-install | Jaeger:kuttl-test-secure-poodle/production-ui updated logger.go:42: 16:13:11 | production/4-install | test step completed 4-install logger.go:42: 16:13:11 | production/5-check-disabled-security | starting test step 5-check-disabled-security logger.go:42: 16:13:11 | production/5-check-disabled-security | running command: [./ensure-ingress-host.sh] logger.go:42: 16:13:11 | production/5-check-disabled-security | Checking the Ingress host value was populated logger.go:42: 16:13:11 | production/5-check-disabled-security | Try number 0 logger.go:42: 16:13:11 | production/5-check-disabled-security | Hostname is production-ui-kuttl-test-secure-poodle.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com logger.go:42: 16:13:11 | production/5-check-disabled-security | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 16:13:12 | production/5-check-disabled-security | Checking an expected HTTP response logger.go:42: 16:13:12 | production/5-check-disabled-security | Running in OpenShift logger.go:42: 16:13:12 | production/5-check-disabled-security | Not using any secret logger.go:42: 16:13:12 | production/5-check-disabled-security | Try number 1/30 the https://production-ui-kuttl-test-secure-poodle.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 16:13:12 | production/5-check-disabled-security | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 16:13:12 | production/5-check-disabled-security | Try number 2/30 the https://production-ui-kuttl-test-secure-poodle.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 16:13:12 | production/5-check-disabled-security | HTTP response is 403. 200 expected. Waiting 10 s logger.go:42: 16:13:22 | production/5-check-disabled-security | Try number 3/30 the https://production-ui-kuttl-test-secure-poodle.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 16:13:22 | production/5-check-disabled-security | curl response asserted properly logger.go:42: 16:13:22 | production/5-check-disabled-security | test step completed 5-check-disabled-security logger.go:42: 16:13:22 | production/6-check-NO-gaID | starting test step 6-check-NO-gaID logger.go:42: 16:13:22 | production/6-check-NO-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 16:13:22 | production/6-check-NO-gaID | Checking the Ingress host value was populated logger.go:42: 16:13:22 | production/6-check-NO-gaID | Try number 0 logger.go:42: 16:13:22 | production/6-check-NO-gaID | Hostname is production-ui-kuttl-test-secure-poodle.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com logger.go:42: 16:13:22 | production/6-check-NO-gaID | running command: [sh -c ASSERT_PRESENT=false EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 16:13:22 | production/6-check-NO-gaID | time="2024-06-17T16:13:22Z" level=info msg="Querying https://production-ui-kuttl-test-secure-poodle.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search..." logger.go:42: 16:13:22 | production/6-check-NO-gaID | time="2024-06-17T16:13:22Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 16:13:22 | production/6-check-NO-gaID | time="2024-06-17T16:13:22Z" level=info msg="Polling to https://production-ui-kuttl-test-secure-poodle.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search" logger.go:42: 16:13:22 | production/6-check-NO-gaID | time="2024-06-17T16:13:22Z" level=info msg="Doing request number 0" logger.go:42: 16:13:22 | production/6-check-NO-gaID | time="2024-06-17T16:13:22Z" level=info msg="Content not found and asserted it was not found!" logger.go:42: 16:13:22 | production/6-check-NO-gaID | time="2024-06-17T16:13:22Z" level=info msg="Success!" logger.go:42: 16:13:22 | production/6-check-NO-gaID | test step completed 6-check-NO-gaID logger.go:42: 16:13:22 | production/7-add-tracking-id | starting test step 7-add-tracking-id logger.go:42: 16:13:22 | production/7-add-tracking-id | running command: [sh -c kubectl apply -f add-tracking-id.yaml -n $NAMESPACE] logger.go:42: 16:13:23 | production/7-add-tracking-id | jaeger.jaegertracing.io/production-ui configured logger.go:42: 16:13:23 | production/7-add-tracking-id | test step completed 7-add-tracking-id logger.go:42: 16:13:23 | production/8-check-gaID | starting test step 8-check-gaID logger.go:42: 16:13:23 | production/8-check-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 16:13:23 | production/8-check-gaID | Checking the Ingress host value was populated logger.go:42: 16:13:23 | production/8-check-gaID | Try number 0 logger.go:42: 16:13:23 | production/8-check-gaID | Hostname is production-ui-kuttl-test-secure-poodle.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com logger.go:42: 16:13:23 | production/8-check-gaID | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 16:13:23 | production/8-check-gaID | time="2024-06-17T16:13:23Z" level=info msg="Querying https://production-ui-kuttl-test-secure-poodle.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search..." logger.go:42: 16:13:23 | production/8-check-gaID | time="2024-06-17T16:13:23Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 16:13:23 | production/8-check-gaID | time="2024-06-17T16:13:23Z" level=info msg="Polling to https://production-ui-kuttl-test-secure-poodle.apps.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com/search" logger.go:42: 16:13:23 | production/8-check-gaID | time="2024-06-17T16:13:23Z" level=info msg="Doing request number 0" logger.go:42: 16:13:23 | production/8-check-gaID | time="2024-06-17T16:13:23Z" level=warning msg="Found: false . Assert: true" logger.go:42: 16:13:23 | production/8-check-gaID | time="2024-06-17T16:13:23Z" level=warning msg="The condition of the test function was not accomplished" logger.go:42: 16:13:31 | production/8-check-gaID | time="2024-06-17T16:13:31Z" level=info msg="Doing request number 1" logger.go:42: 16:13:51 | production/8-check-gaID | time="2024-06-17T16:13:51Z" level=info msg="Content found and asserted!" logger.go:42: 16:13:51 | production/8-check-gaID | time="2024-06-17T16:13:51Z" level=info msg="Success!" logger.go:42: 16:13:52 | production/8-check-gaID | test step completed 8-check-gaID logger.go:42: 16:13:52 | production | production events from ns kuttl-test-secure-poodle: logger.go:42: 16:13:52 | production | 2024-06-17 16:12:23 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestsecurepoodleproductionui-1-df5b5b74 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestsecurepoodleproductionui-1-df5b9lr4w replicaset-controller logger.go:42: 16:13:52 | production | 2024-06-17 16:12:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsecurepoodleproductionui-1-df5b9lr4w Binding Scheduled Successfully assigned kuttl-test-secure-poodle/elasticsearch-cdm-kuttltestsecurepoodleproductionui-1-df5b9lr4w to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:13:52 | production | 2024-06-17 16:12:23 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestsecurepoodleproductionui-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestsecurepoodleproductionui-1-df5b5b74 to 1 deployment-controller logger.go:42: 16:13:52 | production | 2024-06-17 16:12:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsecurepoodleproductionui-1-df5b9lr4w AddedInterface Add eth0 [10.129.2.61/23] from ovn-kubernetes multus logger.go:42: 16:13:52 | production | 2024-06-17 16:12:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsecurepoodleproductionui-1-df5b9lr4w.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsecurepoodleproductionui-1-df5b9lr4w.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsecurepoodleproductionui-1-df5b9lr4w.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsecurepoodleproductionui-1-df5b9lr4w.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsecurepoodleproductionui-1-df5b9lr4w.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsecurepoodleproductionui-1-df5b9lr4w.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:34 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsecurepoodleproductionui-1-df5b9lr4w.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:39 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsecurepoodleproductionui-1-df5b9lr4w.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:50 +0000 UTC Normal Pod production-ui-collector-795c78dc8-2pp7v Binding Scheduled Successfully assigned kuttl-test-secure-poodle/production-ui-collector-795c78dc8-2pp7v to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:13:52 | production | 2024-06-17 16:12:50 +0000 UTC Normal ReplicaSet.apps production-ui-collector-795c78dc8 SuccessfulCreate Created pod: production-ui-collector-795c78dc8-2pp7v replicaset-controller logger.go:42: 16:13:52 | production | 2024-06-17 16:12:50 +0000 UTC Normal Deployment.apps production-ui-collector ScalingReplicaSet Scaled up replica set production-ui-collector-795c78dc8 to 1 deployment-controller logger.go:42: 16:13:52 | production | 2024-06-17 16:12:50 +0000 UTC Normal Pod production-ui-query-6784dff5cf-ttb4f Binding Scheduled Successfully assigned kuttl-test-secure-poodle/production-ui-query-6784dff5cf-ttb4f to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:13:52 | production | 2024-06-17 16:12:50 +0000 UTC Normal ReplicaSet.apps production-ui-query-6784dff5cf SuccessfulCreate Created pod: production-ui-query-6784dff5cf-ttb4f replicaset-controller logger.go:42: 16:13:52 | production | 2024-06-17 16:12:50 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-6784dff5cf to 1 deployment-controller logger.go:42: 16:13:52 | production | 2024-06-17 16:12:51 +0000 UTC Normal Pod production-ui-collector-795c78dc8-2pp7v AddedInterface Add eth0 [10.131.0.79/23] from ovn-kubernetes multus logger.go:42: 16:13:52 | production | 2024-06-17 16:12:51 +0000 UTC Normal Pod production-ui-collector-795c78dc8-2pp7v.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1" already present on machine kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:51 +0000 UTC Normal Pod production-ui-collector-795c78dc8-2pp7v.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:51 +0000 UTC Normal Pod production-ui-collector-795c78dc8-2pp7v.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:51 +0000 UTC Normal Pod production-ui-query-6784dff5cf-ttb4f AddedInterface Add eth0 [10.128.2.50/23] from ovn-kubernetes multus logger.go:42: 16:13:52 | production | 2024-06-17 16:12:51 +0000 UTC Normal Pod production-ui-query-6784dff5cf-ttb4f.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:51 +0000 UTC Normal Pod production-ui-query-6784dff5cf-ttb4f.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:51 +0000 UTC Normal Pod production-ui-query-6784dff5cf-ttb4f.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:51 +0000 UTC Normal Pod production-ui-query-6784dff5cf-ttb4f.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:51 +0000 UTC Normal Pod production-ui-query-6784dff5cf-ttb4f.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:51 +0000 UTC Normal Pod production-ui-query-6784dff5cf-ttb4f.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:51 +0000 UTC Normal Pod production-ui-query-6784dff5cf-ttb4f.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:51 +0000 UTC Normal Pod production-ui-query-6784dff5cf-ttb4f.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:12:51 +0000 UTC Normal Pod production-ui-query-6784dff5cf-ttb4f.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 16:13:52 | production | 2024-06-17 16:13:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 16:13:52 | production | 2024-06-17 16:13:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 16:13:52 | production | 2024-06-17 16:13:06 +0000 UTC Normal Pod production-ui-query-6784dff5cf-ttb4f.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:06 +0000 UTC Normal Pod production-ui-query-6784dff5cf-ttb4f.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:06 +0000 UTC Normal Pod production-ui-query-6784dff5cf-ttb4f.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:06 +0000 UTC Normal ReplicaSet.apps production-ui-query-6784dff5cf SuccessfulDelete Deleted pod: production-ui-query-6784dff5cf-ttb4f replicaset-controller logger.go:42: 16:13:52 | production | 2024-06-17 16:13:06 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-6784dff5cf to 0 from 1 deployment-controller logger.go:42: 16:13:52 | production | 2024-06-17 16:13:07 +0000 UTC Normal Pod production-ui-query-57c7c9cdfb-2rc5v Binding Scheduled Successfully assigned kuttl-test-secure-poodle/production-ui-query-57c7c9cdfb-2rc5v to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:13:52 | production | 2024-06-17 16:13:07 +0000 UTC Normal Pod production-ui-query-57c7c9cdfb-2rc5v AddedInterface Add eth0 [10.128.2.51/23] from ovn-kubernetes multus logger.go:42: 16:13:52 | production | 2024-06-17 16:13:07 +0000 UTC Normal ReplicaSet.apps production-ui-query-57c7c9cdfb SuccessfulCreate Created pod: production-ui-query-57c7c9cdfb-2rc5v replicaset-controller logger.go:42: 16:13:52 | production | 2024-06-17 16:13:07 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-57c7c9cdfb to 1 deployment-controller logger.go:42: 16:13:52 | production | 2024-06-17 16:13:08 +0000 UTC Normal Pod production-ui-query-57c7c9cdfb-2rc5v.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:08 +0000 UTC Normal Pod production-ui-query-57c7c9cdfb-2rc5v.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:08 +0000 UTC Normal Pod production-ui-query-57c7c9cdfb-2rc5v.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:08 +0000 UTC Normal Pod production-ui-query-57c7c9cdfb-2rc5v.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:08 +0000 UTC Normal Pod production-ui-query-57c7c9cdfb-2rc5v.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:08 +0000 UTC Normal Pod production-ui-query-57c7c9cdfb-2rc5v.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:08 +0000 UTC Normal Pod production-ui-query-57c7c9cdfb-2rc5v.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:08 +0000 UTC Normal Pod production-ui-query-57c7c9cdfb-2rc5v.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:08 +0000 UTC Normal Pod production-ui-query-57c7c9cdfb-2rc5v.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:13 +0000 UTC Normal Pod production-ui-query-57c7c9cdfb-2rc5v.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:13 +0000 UTC Normal Pod production-ui-query-57c7c9cdfb-2rc5v.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:13 +0000 UTC Normal Pod production-ui-query-57c7c9cdfb-2rc5v.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:13 +0000 UTC Normal ReplicaSet.apps production-ui-query-57c7c9cdfb SuccessfulDelete Deleted pod: production-ui-query-57c7c9cdfb-2rc5v replicaset-controller logger.go:42: 16:13:52 | production | 2024-06-17 16:13:13 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-57c7c9cdfb to 0 from 1 deployment-controller logger.go:42: 16:13:52 | production | 2024-06-17 16:13:14 +0000 UTC Normal Pod production-ui-query-864f5ff49-7bp56 Binding Scheduled Successfully assigned kuttl-test-secure-poodle/production-ui-query-864f5ff49-7bp56 to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:13:52 | production | 2024-06-17 16:13:14 +0000 UTC Normal ReplicaSet.apps production-ui-query-864f5ff49 SuccessfulCreate Created pod: production-ui-query-864f5ff49-7bp56 replicaset-controller logger.go:42: 16:13:52 | production | 2024-06-17 16:13:14 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-864f5ff49 to 1 deployment-controller logger.go:42: 16:13:52 | production | 2024-06-17 16:13:15 +0000 UTC Normal Pod production-ui-query-864f5ff49-7bp56 AddedInterface Add eth0 [10.128.2.52/23] from ovn-kubernetes multus logger.go:42: 16:13:52 | production | 2024-06-17 16:13:15 +0000 UTC Normal Pod production-ui-query-864f5ff49-7bp56.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:15 +0000 UTC Normal Pod production-ui-query-864f5ff49-7bp56.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:15 +0000 UTC Normal Pod production-ui-query-864f5ff49-7bp56.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:15 +0000 UTC Normal Pod production-ui-query-864f5ff49-7bp56.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:15 +0000 UTC Normal Pod production-ui-query-864f5ff49-7bp56.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:15 +0000 UTC Normal Pod production-ui-query-864f5ff49-7bp56.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:20 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 16:13:52 | production | 2024-06-17 16:13:20 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod production-ui-collector-795c78dc8-2pp7v horizontal-pod-autoscaler logger.go:42: 16:13:52 | production | 2024-06-17 16:13:20 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 16:13:52 | production | 2024-06-17 16:13:24 +0000 UTC Normal Pod production-ui-query-864f5ff49-7bp56.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:24 +0000 UTC Normal Pod production-ui-query-864f5ff49-7bp56.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:24 +0000 UTC Normal ReplicaSet.apps production-ui-query-864f5ff49 SuccessfulDelete Deleted pod: production-ui-query-864f5ff49-7bp56 replicaset-controller logger.go:42: 16:13:52 | production | 2024-06-17 16:13:24 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-864f5ff49 to 0 from 1 deployment-controller logger.go:42: 16:13:52 | production | 2024-06-17 16:13:25 +0000 UTC Normal Pod production-ui-query-54f77d684b-smxhk Binding Scheduled Successfully assigned kuttl-test-secure-poodle/production-ui-query-54f77d684b-smxhk to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:13:52 | production | 2024-06-17 16:13:25 +0000 UTC Normal ReplicaSet.apps production-ui-query-54f77d684b SuccessfulCreate Created pod: production-ui-query-54f77d684b-smxhk replicaset-controller logger.go:42: 16:13:52 | production | 2024-06-17 16:13:25 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-54f77d684b to 1 deployment-controller logger.go:42: 16:13:52 | production | 2024-06-17 16:13:26 +0000 UTC Normal Pod production-ui-query-54f77d684b-smxhk AddedInterface Add eth0 [10.128.2.53/23] from ovn-kubernetes multus logger.go:42: 16:13:52 | production | 2024-06-17 16:13:26 +0000 UTC Normal Pod production-ui-query-54f77d684b-smxhk.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:26 +0000 UTC Normal Pod production-ui-query-54f77d684b-smxhk.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:26 +0000 UTC Normal Pod production-ui-query-54f77d684b-smxhk.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:26 +0000 UTC Normal Pod production-ui-query-54f77d684b-smxhk.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:26 +0000 UTC Normal Pod production-ui-query-54f77d684b-smxhk.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:26 +0000 UTC Normal Pod production-ui-query-54f77d684b-smxhk.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:13:52 | production | 2024-06-17 16:13:35 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod production-ui-collector-795c78dc8-2pp7v horizontal-pod-autoscaler logger.go:42: 16:13:52 | production | 2024-06-17 16:13:35 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod production-ui-collector-795c78dc8-2pp7v horizontal-pod-autoscaler logger.go:42: 16:13:52 | production | Deleting namespace: kuttl-test-secure-poodle === CONT kuttl/harness/artifacts logger.go:42: 16:13:58 | artifacts | Creating namespace: kuttl-test-desired-monster logger.go:42: 16:13:58 | artifacts | artifacts events from ns kuttl-test-desired-monster: logger.go:42: 16:13:58 | artifacts | Deleting namespace: kuttl-test-desired-monster === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (150.48s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/allinone (41.71s) --- PASS: kuttl/harness/production (102.31s) --- PASS: kuttl/harness/artifacts (6.27s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name ui --report --output /logs/artifacts/ui.xml ./artifacts/kuttl-report.xml time="2024-06-17T16:14:05Z" level=debug msg="Setting a new name for the test suites" time="2024-06-17T16:14:05Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-06-17T16:14:05Z" level=debug msg="normalizing test case names" time="2024-06-17T16:14:05Z" level=debug msg="ui/allinone -> ui_allinone" time="2024-06-17T16:14:05Z" level=debug msg="ui/production -> ui_production" time="2024-06-17T16:14:05Z" level=debug msg="ui/artifacts -> ui_artifacts" +---------------+--------+ | NAME | RESULT | +---------------+--------+ | ui_allinone | passed | | ui_production | passed | | ui_artifacts | passed | +---------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh miscellaneous false true + '[' 3 -ne 3 ']' + test_suite_name=miscellaneous + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/miscellaneous.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-miscellaneous make[2]: Entering directory '/tmp/jaeger-tests' SKIP_ES_EXTERNAL=true ./tests/e2e/miscellaneous/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-14-130320 True False 55m Cluster version is 4.16.0-0.nightly-2024-06-14-130320' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-14-130320 True False 55m Cluster version is 4.16.0-0.nightly-2024-06-14-130320' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/miscellaneous/render.sh ++ export SUITE_DIR=./tests/e2e/miscellaneous ++ SUITE_DIR=./tests/e2e/miscellaneous ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/miscellaneous ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test collector-autoscale + '[' 1 -ne 1 ']' + test_name=collector-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-autoscale\e[0m' Rendering files for test collector-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p collector-autoscale + cd collector-autoscale + jaeger_name=simple-prod + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + ELASTICSEARCH_NODECOUNT=1 + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.resources.requests.memory="200m"' 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.autoscale=true 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.minReplicas=1 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.maxReplicas=2 01-install.yaml + version_lt 1.27 1.23 ++ echo 1.27 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.27 '!=' 1.27 + rm ./03-assert.yaml + generate_otlp_e2e_tests http + test_protocol=http + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-http\e[0m' Rendering files for test collector-otlp-allinone-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-autoscale + '[' collector-autoscale '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-http + cd collector-otlp-allinone-http + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger http true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-http\e[0m' Rendering files for test collector-otlp-production-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-http + '[' collector-otlp-allinone-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-http + cd collector-otlp-production-http + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger http true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + generate_otlp_e2e_tests grpc + test_protocol=grpc + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-grpc\e[0m' Rendering files for test collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-http + '[' collector-otlp-production-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-grpc + cd collector-otlp-allinone-grpc + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-grpc\e[0m' Rendering files for test collector-otlp-production-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-grpc + '[' collector-otlp-allinone-grpc '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-grpc + cd collector-otlp-production-grpc + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + '[' true = true ']' + skip_test istio 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=istio + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-grpc + '[' collector-otlp-production-grpc '!=' _build ']' + cd .. + rm -rf istio + warning 'istio: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: istio: Test not supported in OpenShift\e[0m' WAR: istio: Test not supported in OpenShift + '[' true = true ']' + skip_test outside-cluster 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=outside-cluster + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf outside-cluster + warning 'outside-cluster: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: outside-cluster: Test not supported in OpenShift\e[0m' WAR: outside-cluster: Test not supported in OpenShift + start_test set-custom-img + '[' 1 -ne 1 ']' + test_name=set-custom-img + echo =========================================================================== =========================================================================== + info 'Rendering files for test set-custom-img' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test set-custom-img\e[0m' Rendering files for test set-custom-img + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p set-custom-img + cd set-custom-img + jaeger_name=my-jaeger + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.image="test"' ./02-install.yaml + '[' true = true ']' + skip_test non-cluster-wide 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=non-cluster-wide + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/set-custom-img + '[' set-custom-img '!=' _build ']' + cd .. + rm -rf non-cluster-wide + warning 'non-cluster-wide: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: non-cluster-wide: Test not supported in OpenShift\e[0m' WAR: non-cluster-wide: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running miscellaneous E2E tests' Running miscellaneous E2E tests + cd tests/e2e/miscellaneous/_build + set +e + KUBECONFIG=/tmp/kubeconfig-37736524 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-29jb8lcj-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 8 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/cassandra-spark === PAUSE kuttl/harness/cassandra-spark === RUN kuttl/harness/collector-autoscale === PAUSE kuttl/harness/collector-autoscale === RUN kuttl/harness/collector-otlp-allinone-grpc === PAUSE kuttl/harness/collector-otlp-allinone-grpc === RUN kuttl/harness/collector-otlp-allinone-http === PAUSE kuttl/harness/collector-otlp-allinone-http === RUN kuttl/harness/collector-otlp-production-grpc === PAUSE kuttl/harness/collector-otlp-production-grpc === RUN kuttl/harness/collector-otlp-production-http === PAUSE kuttl/harness/collector-otlp-production-http === RUN kuttl/harness/set-custom-img === PAUSE kuttl/harness/set-custom-img === CONT kuttl/harness/artifacts logger.go:42: 16:14:16 | artifacts | Creating namespace: kuttl-test-communal-katydid logger.go:42: 16:14:16 | artifacts | artifacts events from ns kuttl-test-communal-katydid: logger.go:42: 16:14:16 | artifacts | Deleting namespace: kuttl-test-communal-katydid === CONT kuttl/harness/collector-otlp-allinone-http logger.go:42: 16:14:22 | collector-otlp-allinone-http | Creating namespace: kuttl-test-enabled-anchovy logger.go:42: 16:14:22 | collector-otlp-allinone-http/0-install | starting test step 0-install logger.go:42: 16:14:22 | collector-otlp-allinone-http/0-install | Jaeger:kuttl-test-enabled-anchovy/my-jaeger created logger.go:42: 16:14:27 | collector-otlp-allinone-http/0-install | test step completed 0-install logger.go:42: 16:14:27 | collector-otlp-allinone-http/1-smoke-test | starting test step 1-smoke-test logger.go:42: 16:14:27 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 16:14:29 | collector-otlp-allinone-http/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 16:14:35 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 16:14:36 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 16:14:36 | collector-otlp-allinone-http/1-smoke-test | job.batch/report-span created logger.go:42: 16:14:36 | collector-otlp-allinone-http/1-smoke-test | job.batch/check-span created logger.go:42: 16:14:48 | collector-otlp-allinone-http/1-smoke-test | test step completed 1-smoke-test logger.go:42: 16:14:48 | collector-otlp-allinone-http | collector-otlp-allinone-http events from ns kuttl-test-enabled-anchovy: logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:26 +0000 UTC Normal Pod my-jaeger-6b9ffd59b6-l4fgm Binding Scheduled Successfully assigned kuttl-test-enabled-anchovy/my-jaeger-6b9ffd59b6-l4fgm to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:26 +0000 UTC Normal Pod my-jaeger-6b9ffd59b6-l4fgm AddedInterface Add eth0 [10.129.2.62/23] from ovn-kubernetes multus logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:26 +0000 UTC Normal Pod my-jaeger-6b9ffd59b6-l4fgm.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:26 +0000 UTC Normal Pod my-jaeger-6b9ffd59b6-l4fgm.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:26 +0000 UTC Normal Pod my-jaeger-6b9ffd59b6-l4fgm.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:26 +0000 UTC Normal Pod my-jaeger-6b9ffd59b6-l4fgm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:26 +0000 UTC Normal ReplicaSet.apps my-jaeger-6b9ffd59b6 SuccessfulCreate Created pod: my-jaeger-6b9ffd59b6-l4fgm replicaset-controller logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:26 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6b9ffd59b6 to 1 deployment-controller logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:27 +0000 UTC Normal Pod my-jaeger-6b9ffd59b6-l4fgm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:27 +0000 UTC Normal Pod my-jaeger-6b9ffd59b6-l4fgm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:31 +0000 UTC Normal Pod my-jaeger-6b9ffd59b6-l4fgm.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:31 +0000 UTC Normal Pod my-jaeger-6b9ffd59b6-l4fgm.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:31 +0000 UTC Normal ReplicaSet.apps my-jaeger-6b9ffd59b6 SuccessfulDelete Deleted pod: my-jaeger-6b9ffd59b6-l4fgm replicaset-controller logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:31 +0000 UTC Normal Pod my-jaeger-79bd4fbb7c-xqwrr Binding Scheduled Successfully assigned kuttl-test-enabled-anchovy/my-jaeger-79bd4fbb7c-xqwrr to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:31 +0000 UTC Normal ReplicaSet.apps my-jaeger-79bd4fbb7c SuccessfulCreate Created pod: my-jaeger-79bd4fbb7c-xqwrr replicaset-controller logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:31 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-6b9ffd59b6 to 0 from 1 deployment-controller logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:31 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-79bd4fbb7c to 1 deployment-controller logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:32 +0000 UTC Normal Pod my-jaeger-79bd4fbb7c-xqwrr AddedInterface Add eth0 [10.129.2.63/23] from ovn-kubernetes multus logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:32 +0000 UTC Normal Pod my-jaeger-79bd4fbb7c-xqwrr.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:32 +0000 UTC Normal Pod my-jaeger-79bd4fbb7c-xqwrr.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:32 +0000 UTC Normal Pod my-jaeger-79bd4fbb7c-xqwrr.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:32 +0000 UTC Normal Pod my-jaeger-79bd4fbb7c-xqwrr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:32 +0000 UTC Normal Pod my-jaeger-79bd4fbb7c-xqwrr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:32 +0000 UTC Normal Pod my-jaeger-79bd4fbb7c-xqwrr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:36 +0000 UTC Normal Pod check-span-vzmxf Binding Scheduled Successfully assigned kuttl-test-enabled-anchovy/check-span-vzmxf to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:36 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-vzmxf job-controller logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:36 +0000 UTC Normal Pod report-span-tl94v Binding Scheduled Successfully assigned kuttl-test-enabled-anchovy/report-span-tl94v to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:36 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-tl94v job-controller logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:37 +0000 UTC Normal Pod check-span-vzmxf AddedInterface Add eth0 [10.128.2.54/23] from ovn-kubernetes multus logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:37 +0000 UTC Normal Pod check-span-vzmxf.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:37 +0000 UTC Normal Pod check-span-vzmxf.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 401ms (401ms including waiting) kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:37 +0000 UTC Normal Pod check-span-vzmxf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:37 +0000 UTC Normal Pod check-span-vzmxf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:37 +0000 UTC Normal Pod report-span-tl94v AddedInterface Add eth0 [10.131.0.80/23] from ovn-kubernetes multus logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:37 +0000 UTC Normal Pod report-span-tl94v.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:37 +0000 UTC Normal Pod report-span-tl94v.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 429ms (429ms including waiting) kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:37 +0000 UTC Normal Pod report-span-tl94v.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:37 +0000 UTC Normal Pod report-span-tl94v.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 16:14:48 | collector-otlp-allinone-http | 2024-06-17 16:14:48 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 16:14:48 | collector-otlp-allinone-http | Deleting namespace: kuttl-test-enabled-anchovy === CONT kuttl/harness/set-custom-img logger.go:42: 16:14:56 | set-custom-img | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 16:14:56 | set-custom-img | Ignoring check-collector-img.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 16:14:56 | set-custom-img | Creating namespace: kuttl-test-helped-longhorn logger.go:42: 16:14:56 | set-custom-img/1-install | starting test step 1-install logger.go:42: 16:14:56 | set-custom-img/1-install | Jaeger:kuttl-test-helped-longhorn/my-jaeger created logger.go:42: 16:15:32 | set-custom-img/1-install | test step completed 1-install logger.go:42: 16:15:32 | set-custom-img/2-install | starting test step 2-install logger.go:42: 16:15:33 | set-custom-img/2-install | Jaeger:kuttl-test-helped-longhorn/my-jaeger updated logger.go:42: 16:15:33 | set-custom-img/2-install | test step completed 2-install logger.go:42: 16:15:33 | set-custom-img/3-check-image | starting test step 3-check-image logger.go:42: 16:15:33 | set-custom-img/3-check-image | running command: [sh -c ./check-collector-img.sh] logger.go:42: 16:15:33 | set-custom-img/3-check-image | Collector image missmatch. Expected: test. Has: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1 logger.go:42: 16:15:38 | set-custom-img/3-check-image | Collector image asserted properly! logger.go:42: 16:15:38 | set-custom-img/3-check-image | test step completed 3-check-image logger.go:42: 16:15:38 | set-custom-img | set-custom-img events from ns kuttl-test-helped-longhorn: logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthelpedlonghornmyjaeger-1-69978c6hdms Binding Scheduled Successfully assigned kuttl-test-helped-longhorn/elasticsearch-cdm-kuttltesthelpedlonghornmyjaeger-1-69978c6hdms to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:02 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltesthelpedlonghornmyjaeger-1-69978c7c5d SuccessfulCreate Created pod: elasticsearch-cdm-kuttltesthelpedlonghornmyjaeger-1-69978c6hdms replicaset-controller logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:02 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltesthelpedlonghornmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltesthelpedlonghornmyjaeger-1-69978c7c5d to 1 deployment-controller logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthelpedlonghornmyjaeger-1-69978c6hdms AddedInterface Add eth0 [10.129.2.65/23] from ovn-kubernetes multus logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthelpedlonghornmyjaeger-1-69978c6hdms.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthelpedlonghornmyjaeger-1-69978c6hdms.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthelpedlonghornmyjaeger-1-69978c6hdms.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthelpedlonghornmyjaeger-1-69978c6hdms.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthelpedlonghornmyjaeger-1-69978c6hdms.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthelpedlonghornmyjaeger-1-69978c6hdms.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:18 +0000 UTC Warning Pod elasticsearch-cdm-kuttltesthelpedlonghornmyjaeger-1-69978c6hdms.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:29 +0000 UTC Normal Pod my-jaeger-collector-6fc4f5bcf9-bjkm7 Binding Scheduled Successfully assigned kuttl-test-helped-longhorn/my-jaeger-collector-6fc4f5bcf9-bjkm7 to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:29 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-6fc4f5bcf9 SuccessfulCreate Created pod: my-jaeger-collector-6fc4f5bcf9-bjkm7 replicaset-controller logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:29 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-6fc4f5bcf9 to 1 deployment-controller logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:29 +0000 UTC Normal Pod my-jaeger-query-9c4db8458-ztrfc Binding Scheduled Successfully assigned kuttl-test-helped-longhorn/my-jaeger-query-9c4db8458-ztrfc to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:29 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-9c4db8458 SuccessfulCreate Created pod: my-jaeger-query-9c4db8458-ztrfc replicaset-controller logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:29 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-9c4db8458 to 1 deployment-controller logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:30 +0000 UTC Normal Pod my-jaeger-collector-6fc4f5bcf9-bjkm7 AddedInterface Add eth0 [10.131.0.81/23] from ovn-kubernetes multus logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:30 +0000 UTC Normal Pod my-jaeger-collector-6fc4f5bcf9-bjkm7.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1" already present on machine kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:30 +0000 UTC Normal Pod my-jaeger-collector-6fc4f5bcf9-bjkm7.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:30 +0000 UTC Normal Pod my-jaeger-collector-6fc4f5bcf9-bjkm7.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:30 +0000 UTC Normal Pod my-jaeger-query-9c4db8458-ztrfc AddedInterface Add eth0 [10.128.2.55/23] from ovn-kubernetes multus logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:30 +0000 UTC Normal Pod my-jaeger-query-9c4db8458-ztrfc.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:30 +0000 UTC Normal Pod my-jaeger-query-9c4db8458-ztrfc.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:30 +0000 UTC Normal Pod my-jaeger-query-9c4db8458-ztrfc.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:30 +0000 UTC Normal Pod my-jaeger-query-9c4db8458-ztrfc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:30 +0000 UTC Normal Pod my-jaeger-query-9c4db8458-ztrfc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:30 +0000 UTC Normal Pod my-jaeger-query-9c4db8458-ztrfc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:30 +0000 UTC Normal Pod my-jaeger-query-9c4db8458-ztrfc.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:30 +0000 UTC Normal Pod my-jaeger-query-9c4db8458-ztrfc.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:30 +0000 UTC Normal Pod my-jaeger-query-9c4db8458-ztrfc.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:36 +0000 UTC Normal Pod my-jaeger-collector-6fc4f5bcf9-bjkm7.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:36 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-6fc4f5bcf9 SuccessfulDelete Deleted pod: my-jaeger-collector-6fc4f5bcf9-bjkm7 replicaset-controller logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:36 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-6fc4f5bcf9 to 0 from 1 deployment-controller logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:37 +0000 UTC Normal Pod my-jaeger-collector-7d89b548cc-k868z Binding Scheduled Successfully assigned kuttl-test-helped-longhorn/my-jaeger-collector-7d89b548cc-k868z to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:37 +0000 UTC Normal Pod my-jaeger-collector-7d89b548cc-k868z AddedInterface Add eth0 [10.131.0.82/23] from ovn-kubernetes multus logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:37 +0000 UTC Normal Pod my-jaeger-collector-7d89b548cc-k868z.spec.containers{jaeger-collector} Pulling Pulling image "test" kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:37 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7d89b548cc SuccessfulCreate Created pod: my-jaeger-collector-7d89b548cc-k868z replicaset-controller logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:37 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7d89b548cc to 1 deployment-controller logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:38 +0000 UTC Warning Pod my-jaeger-collector-7d89b548cc-k868z.spec.containers{jaeger-collector} Failed Failed to pull image "test": reading manifest latest in docker.io/library/test: requested access to the resource is denied kubelet logger.go:42: 16:15:38 | set-custom-img | 2024-06-17 16:15:38 +0000 UTC Warning Pod my-jaeger-collector-7d89b548cc-k868z.spec.containers{jaeger-collector} Failed Error: ErrImagePull kubelet logger.go:42: 16:15:38 | set-custom-img | Deleting namespace: kuttl-test-helped-longhorn === CONT kuttl/harness/collector-otlp-production-http logger.go:42: 16:15:44 | collector-otlp-production-http | Creating namespace: kuttl-test-glowing-pegasus logger.go:42: 16:15:44 | collector-otlp-production-http/1-install | starting test step 1-install logger.go:42: 16:15:45 | collector-otlp-production-http/1-install | Jaeger:kuttl-test-glowing-pegasus/my-jaeger created logger.go:42: 16:16:21 | collector-otlp-production-http/1-install | test step completed 1-install logger.go:42: 16:16:21 | collector-otlp-production-http/2-smoke-test | starting test step 2-smoke-test logger.go:42: 16:16:21 | collector-otlp-production-http/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 16:16:22 | collector-otlp-production-http/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 16:16:29 | collector-otlp-production-http/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 16:16:29 | collector-otlp-production-http/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 16:16:30 | collector-otlp-production-http/2-smoke-test | job.batch/report-span created logger.go:42: 16:16:30 | collector-otlp-production-http/2-smoke-test | job.batch/check-span created logger.go:42: 16:16:42 | collector-otlp-production-http/2-smoke-test | test step completed 2-smoke-test logger.go:42: 16:16:42 | collector-otlp-production-http | collector-otlp-production-http events from ns kuttl-test-glowing-pegasus: logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:15:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestglowingpegasusmyjaeger-1-59f89b822ph Binding Scheduled Successfully assigned kuttl-test-glowing-pegasus/elasticsearch-cdm-kuttltestglowingpegasusmyjaeger-1-59f89b822ph to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:15:50 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestglowingpegasusmyjaeger-1-59f89bdc78 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestglowingpegasusmyjaeger-1-59f89b822ph replicaset-controller logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:15:50 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestglowingpegasusmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestglowingpegasusmyjaeger-1-59f89bdc78 to 1 deployment-controller logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:15:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestglowingpegasusmyjaeger-1-59f89b822ph AddedInterface Add eth0 [10.129.2.66/23] from ovn-kubernetes multus logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:15:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestglowingpegasusmyjaeger-1-59f89b822ph.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:15:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestglowingpegasusmyjaeger-1-59f89b822ph.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:15:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestglowingpegasusmyjaeger-1-59f89b822ph.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:15:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestglowingpegasusmyjaeger-1-59f89b822ph.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:15:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestglowingpegasusmyjaeger-1-59f89b822ph.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:15:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestglowingpegasusmyjaeger-1-59f89b822ph.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:01 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestglowingpegasusmyjaeger-1-59f89b822ph.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:06 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestglowingpegasusmyjaeger-1-59f89b822ph.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:17 +0000 UTC Normal Pod my-jaeger-collector-6fccf5d447-f4l8z Binding Scheduled Successfully assigned kuttl-test-glowing-pegasus/my-jaeger-collector-6fccf5d447-f4l8z to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:17 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-6fccf5d447 SuccessfulCreate Created pod: my-jaeger-collector-6fccf5d447-f4l8z replicaset-controller logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:17 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-6fccf5d447 to 1 deployment-controller logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:17 +0000 UTC Normal Pod my-jaeger-query-6c56ff7dd7-mtmcf Binding Scheduled Successfully assigned kuttl-test-glowing-pegasus/my-jaeger-query-6c56ff7dd7-mtmcf to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:17 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-6c56ff7dd7 SuccessfulCreate Created pod: my-jaeger-query-6c56ff7dd7-mtmcf replicaset-controller logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:17 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-6c56ff7dd7 to 1 deployment-controller logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:18 +0000 UTC Normal Pod my-jaeger-collector-6fccf5d447-f4l8z AddedInterface Add eth0 [10.131.0.83/23] from ovn-kubernetes multus logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:18 +0000 UTC Normal Pod my-jaeger-collector-6fccf5d447-f4l8z.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1" already present on machine kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:18 +0000 UTC Normal Pod my-jaeger-collector-6fccf5d447-f4l8z.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:18 +0000 UTC Normal Pod my-jaeger-collector-6fccf5d447-f4l8z.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:18 +0000 UTC Normal Pod my-jaeger-query-6c56ff7dd7-mtmcf AddedInterface Add eth0 [10.128.2.56/23] from ovn-kubernetes multus logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:18 +0000 UTC Normal Pod my-jaeger-query-6c56ff7dd7-mtmcf.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:18 +0000 UTC Normal Pod my-jaeger-query-6c56ff7dd7-mtmcf.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:18 +0000 UTC Normal Pod my-jaeger-query-6c56ff7dd7-mtmcf.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:18 +0000 UTC Normal Pod my-jaeger-query-6c56ff7dd7-mtmcf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:18 +0000 UTC Normal Pod my-jaeger-query-6c56ff7dd7-mtmcf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:18 +0000 UTC Normal Pod my-jaeger-query-6c56ff7dd7-mtmcf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:18 +0000 UTC Normal Pod my-jaeger-query-6c56ff7dd7-mtmcf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:18 +0000 UTC Normal Pod my-jaeger-query-6c56ff7dd7-mtmcf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:18 +0000 UTC Normal Pod my-jaeger-query-6c56ff7dd7-mtmcf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:24 +0000 UTC Normal Pod my-jaeger-query-68c7db6689-457b2 Binding Scheduled Successfully assigned kuttl-test-glowing-pegasus/my-jaeger-query-68c7db6689-457b2 to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:24 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-68c7db6689 SuccessfulCreate Created pod: my-jaeger-query-68c7db6689-457b2 replicaset-controller logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:24 +0000 UTC Normal Pod my-jaeger-query-6c56ff7dd7-mtmcf.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:24 +0000 UTC Normal Pod my-jaeger-query-6c56ff7dd7-mtmcf.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:24 +0000 UTC Normal Pod my-jaeger-query-6c56ff7dd7-mtmcf.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:24 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-6c56ff7dd7 SuccessfulDelete Deleted pod: my-jaeger-query-6c56ff7dd7-mtmcf replicaset-controller logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:24 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-6c56ff7dd7 to 0 from 1 deployment-controller logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:24 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-68c7db6689 to 1 deployment-controller logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:25 +0000 UTC Normal Pod my-jaeger-query-68c7db6689-457b2 AddedInterface Add eth0 [10.128.2.57/23] from ovn-kubernetes multus logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:25 +0000 UTC Normal Pod my-jaeger-query-68c7db6689-457b2.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:25 +0000 UTC Normal Pod my-jaeger-query-68c7db6689-457b2.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:25 +0000 UTC Normal Pod my-jaeger-query-68c7db6689-457b2.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:25 +0000 UTC Normal Pod my-jaeger-query-68c7db6689-457b2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:25 +0000 UTC Normal Pod my-jaeger-query-68c7db6689-457b2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:25 +0000 UTC Normal Pod my-jaeger-query-68c7db6689-457b2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:25 +0000 UTC Normal Pod my-jaeger-query-68c7db6689-457b2.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:25 +0000 UTC Normal Pod my-jaeger-query-68c7db6689-457b2.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:25 +0000 UTC Normal Pod my-jaeger-query-68c7db6689-457b2.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:30 +0000 UTC Normal Pod check-span-67r8w Binding Scheduled Successfully assigned kuttl-test-glowing-pegasus/check-span-67r8w to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:30 +0000 UTC Normal Pod check-span-67r8w AddedInterface Add eth0 [10.131.0.85/23] from ovn-kubernetes multus logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:30 +0000 UTC Normal Pod check-span-67r8w.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:30 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-67r8w job-controller logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:30 +0000 UTC Normal Pod report-span-fkc8p Binding Scheduled Successfully assigned kuttl-test-glowing-pegasus/report-span-fkc8p to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:30 +0000 UTC Normal Pod report-span-fkc8p AddedInterface Add eth0 [10.131.0.84/23] from ovn-kubernetes multus logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:30 +0000 UTC Normal Pod report-span-fkc8p.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:30 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-fkc8p job-controller logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:31 +0000 UTC Normal Pod check-span-67r8w.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 394ms (394ms including waiting) kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:31 +0000 UTC Normal Pod check-span-67r8w.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:31 +0000 UTC Normal Pod check-span-67r8w.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:32 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:32 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:32 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:32 +0000 UTC Normal Pod report-span-fkc8p.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.473s (1.473s including waiting) kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:32 +0000 UTC Normal Pod report-span-fkc8p.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:32 +0000 UTC Normal Pod report-span-fkc8p.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 16:16:42 | collector-otlp-production-http | 2024-06-17 16:16:42 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 16:16:42 | collector-otlp-production-http | Deleting namespace: kuttl-test-glowing-pegasus === CONT kuttl/harness/collector-otlp-production-grpc logger.go:42: 16:16:55 | collector-otlp-production-grpc | Creating namespace: kuttl-test-splendid-starfish logger.go:42: 16:16:55 | collector-otlp-production-grpc/1-install | starting test step 1-install logger.go:42: 16:16:55 | collector-otlp-production-grpc/1-install | Jaeger:kuttl-test-splendid-starfish/my-jaeger created logger.go:42: 16:17:31 | collector-otlp-production-grpc/1-install | test step completed 1-install logger.go:42: 16:17:31 | collector-otlp-production-grpc/2-smoke-test | starting test step 2-smoke-test logger.go:42: 16:17:31 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 16:17:33 | collector-otlp-production-grpc/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 16:17:39 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 16:17:39 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 16:17:40 | collector-otlp-production-grpc/2-smoke-test | job.batch/report-span created logger.go:42: 16:17:40 | collector-otlp-production-grpc/2-smoke-test | job.batch/check-span created logger.go:42: 16:18:00 | collector-otlp-production-grpc/2-smoke-test | test step completed 2-smoke-test logger.go:42: 16:18:00 | collector-otlp-production-grpc | collector-otlp-production-grpc events from ns kuttl-test-splendid-starfish: logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:01 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsplendidstarfishmyjaeger-1-db6b54ccw Binding Scheduled Successfully assigned kuttl-test-splendid-starfish/elasticsearch-cdm-kuttltestsplendidstarfishmyjaeger-1-db6b54ccw to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:01 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestsplendidstarfishmyjaeger-1-db6b9b5c8 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestsplendidstarfishmyjaeger-1-db6b54ccw replicaset-controller logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:01 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestsplendidstarfishmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestsplendidstarfishmyjaeger-1-db6b9b5c8 to 1 deployment-controller logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsplendidstarfishmyjaeger-1-db6b54ccw AddedInterface Add eth0 [10.129.2.67/23] from ovn-kubernetes multus logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsplendidstarfishmyjaeger-1-db6b54ccw.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsplendidstarfishmyjaeger-1-db6b54ccw.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsplendidstarfishmyjaeger-1-db6b54ccw.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsplendidstarfishmyjaeger-1-db6b54ccw.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsplendidstarfishmyjaeger-1-db6b54ccw.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsplendidstarfishmyjaeger-1-db6b54ccw.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:12 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsplendidstarfishmyjaeger-1-db6b54ccw.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:17 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsplendidstarfishmyjaeger-1-db6b54ccw.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:28 +0000 UTC Normal Pod my-jaeger-collector-77bb549689-pgm72 Binding Scheduled Successfully assigned kuttl-test-splendid-starfish/my-jaeger-collector-77bb549689-pgm72 to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:28 +0000 UTC Warning Pod my-jaeger-collector-77bb549689-pgm72 FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:28 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-77bb549689 SuccessfulCreate Created pod: my-jaeger-collector-77bb549689-pgm72 replicaset-controller logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:28 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-77bb549689 to 1 deployment-controller logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:28 +0000 UTC Normal Pod my-jaeger-query-6c9956774c-cfl8p Binding Scheduled Successfully assigned kuttl-test-splendid-starfish/my-jaeger-query-6c9956774c-cfl8p to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:28 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-6c9956774c SuccessfulCreate Created pod: my-jaeger-query-6c9956774c-cfl8p replicaset-controller logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:28 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-6c9956774c to 1 deployment-controller logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:29 +0000 UTC Normal Pod my-jaeger-collector-77bb549689-pgm72 AddedInterface Add eth0 [10.131.0.86/23] from ovn-kubernetes multus logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:29 +0000 UTC Normal Pod my-jaeger-collector-77bb549689-pgm72.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1" already present on machine kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:29 +0000 UTC Normal Pod my-jaeger-query-6c9956774c-cfl8p AddedInterface Add eth0 [10.128.2.58/23] from ovn-kubernetes multus logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:29 +0000 UTC Normal Pod my-jaeger-query-6c9956774c-cfl8p.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:29 +0000 UTC Normal Pod my-jaeger-query-6c9956774c-cfl8p.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:29 +0000 UTC Normal Pod my-jaeger-query-6c9956774c-cfl8p.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:29 +0000 UTC Normal Pod my-jaeger-query-6c9956774c-cfl8p.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:29 +0000 UTC Normal Pod my-jaeger-query-6c9956774c-cfl8p.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:29 +0000 UTC Normal Pod my-jaeger-query-6c9956774c-cfl8p.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:29 +0000 UTC Normal Pod my-jaeger-query-6c9956774c-cfl8p.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:29 +0000 UTC Normal Pod my-jaeger-query-6c9956774c-cfl8p.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:29 +0000 UTC Normal Pod my-jaeger-query-6c9956774c-cfl8p.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:30 +0000 UTC Normal Pod my-jaeger-collector-77bb549689-pgm72.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:30 +0000 UTC Normal Pod my-jaeger-collector-77bb549689-pgm72.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:34 +0000 UTC Normal Pod my-jaeger-query-6c9956774c-cfl8p.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:34 +0000 UTC Normal Pod my-jaeger-query-6c9956774c-cfl8p.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:34 +0000 UTC Normal Pod my-jaeger-query-6c9956774c-cfl8p.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:34 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-6c9956774c SuccessfulDelete Deleted pod: my-jaeger-query-6c9956774c-cfl8p replicaset-controller logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:34 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-6c9956774c to 0 from 1 deployment-controller logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:35 +0000 UTC Normal Pod my-jaeger-query-647c6db87f-dbzgl Binding Scheduled Successfully assigned kuttl-test-splendid-starfish/my-jaeger-query-647c6db87f-dbzgl to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:35 +0000 UTC Normal Pod my-jaeger-query-647c6db87f-dbzgl AddedInterface Add eth0 [10.128.2.59/23] from ovn-kubernetes multus logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:35 +0000 UTC Normal Pod my-jaeger-query-647c6db87f-dbzgl.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:35 +0000 UTC Normal Pod my-jaeger-query-647c6db87f-dbzgl.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:35 +0000 UTC Normal Pod my-jaeger-query-647c6db87f-dbzgl.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:35 +0000 UTC Normal Pod my-jaeger-query-647c6db87f-dbzgl.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:35 +0000 UTC Normal Pod my-jaeger-query-647c6db87f-dbzgl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:35 +0000 UTC Normal Pod my-jaeger-query-647c6db87f-dbzgl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:35 +0000 UTC Normal Pod my-jaeger-query-647c6db87f-dbzgl.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:35 +0000 UTC Normal Pod my-jaeger-query-647c6db87f-dbzgl.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:35 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-647c6db87f SuccessfulCreate Created pod: my-jaeger-query-647c6db87f-dbzgl replicaset-controller logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:35 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-647c6db87f to 1 deployment-controller logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:36 +0000 UTC Normal Pod my-jaeger-query-647c6db87f-dbzgl.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:40 +0000 UTC Normal Pod check-span-nrgw9 Binding Scheduled Successfully assigned kuttl-test-splendid-starfish/check-span-nrgw9 to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:40 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-nrgw9 job-controller logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:40 +0000 UTC Normal Pod report-span-f25rp Binding Scheduled Successfully assigned kuttl-test-splendid-starfish/report-span-f25rp to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:40 +0000 UTC Normal Pod report-span-f25rp AddedInterface Add eth0 [10.131.0.87/23] from ovn-kubernetes multus logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:40 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-f25rp job-controller logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:41 +0000 UTC Normal Pod check-span-nrgw9 AddedInterface Add eth0 [10.131.0.88/23] from ovn-kubernetes multus logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:41 +0000 UTC Normal Pod check-span-nrgw9.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:41 +0000 UTC Normal Pod check-span-nrgw9.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 317ms (317ms including waiting) kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:41 +0000 UTC Normal Pod check-span-nrgw9.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:41 +0000 UTC Normal Pod check-span-nrgw9.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:41 +0000 UTC Normal Pod report-span-f25rp.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:41 +0000 UTC Normal Pod report-span-f25rp.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 397ms (397ms including waiting) kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:41 +0000 UTC Normal Pod report-span-f25rp.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:41 +0000 UTC Normal Pod report-span-f25rp.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:17:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 16:18:00 | collector-otlp-production-grpc | 2024-06-17 16:18:00 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 16:18:00 | collector-otlp-production-grpc | Deleting namespace: kuttl-test-splendid-starfish === CONT kuttl/harness/collector-autoscale logger.go:42: 16:18:13 | collector-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 16:18:13 | collector-autoscale | Creating namespace: kuttl-test-handy-oyster logger.go:42: 16:18:13 | collector-autoscale/1-install | starting test step 1-install logger.go:42: 16:18:13 | collector-autoscale/1-install | Jaeger:kuttl-test-handy-oyster/simple-prod created logger.go:42: 16:18:50 | collector-autoscale/1-install | test step completed 1-install logger.go:42: 16:18:50 | collector-autoscale/2- | starting test step 2- logger.go:42: 16:18:52 | collector-autoscale/2- | test step completed 2- logger.go:42: 16:18:52 | collector-autoscale | collector-autoscale events from ns kuttl-test-handy-oyster: logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:20 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltesthandyoystersimpleprod-1-5d95c456cd SuccessfulCreate Created pod: elasticsearch-cdm-kuttltesthandyoystersimpleprod-1-5d95c45rbjws replicaset-controller logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthandyoystersimpleprod-1-5d95c45rbjws Binding Scheduled Successfully assigned kuttl-test-handy-oyster/elasticsearch-cdm-kuttltesthandyoystersimpleprod-1-5d95c45rbjws to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:20 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltesthandyoystersimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltesthandyoystersimpleprod-1-5d95c456cd to 1 deployment-controller logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthandyoystersimpleprod-1-5d95c45rbjws AddedInterface Add eth0 [10.129.2.68/23] from ovn-kubernetes multus logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthandyoystersimpleprod-1-5d95c45rbjws.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthandyoystersimpleprod-1-5d95c45rbjws.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthandyoystersimpleprod-1-5d95c45rbjws.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthandyoystersimpleprod-1-5d95c45rbjws.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthandyoystersimpleprod-1-5d95c45rbjws.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthandyoystersimpleprod-1-5d95c45rbjws.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:36 +0000 UTC Warning Pod elasticsearch-cdm-kuttltesthandyoystersimpleprod-1-5d95c45rbjws.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:47 +0000 UTC Normal Pod simple-prod-collector-586d5b59d8-j7r96 Binding Scheduled Successfully assigned kuttl-test-handy-oyster/simple-prod-collector-586d5b59d8-j7r96 to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:47 +0000 UTC Normal Pod simple-prod-collector-586d5b59d8-j7r96 AddedInterface Add eth0 [10.131.0.89/23] from ovn-kubernetes multus logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:47 +0000 UTC Normal Pod simple-prod-collector-586d5b59d8-j7r96.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:02acff0dd99059a06a7a121c87952b4f70f835e456cb68174ea4cd913e9f8da1" already present on machine kubelet logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:47 +0000 UTC Normal Pod simple-prod-collector-586d5b59d8-j7r96.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:47 +0000 UTC Normal Pod simple-prod-collector-586d5b59d8-j7r96.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:47 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-586d5b59d8 SuccessfulCreate Created pod: simple-prod-collector-586d5b59d8-j7r96 replicaset-controller logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:47 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-586d5b59d8 to 1 deployment-controller logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:47 +0000 UTC Normal Pod simple-prod-query-755b7d8989-nr9m7 Binding Scheduled Successfully assigned kuttl-test-handy-oyster/simple-prod-query-755b7d8989-nr9m7 to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:47 +0000 UTC Normal Pod simple-prod-query-755b7d8989-nr9m7 AddedInterface Add eth0 [10.128.2.60/23] from ovn-kubernetes multus logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:47 +0000 UTC Normal Pod simple-prod-query-755b7d8989-nr9m7.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:e99378a6bedc68c412b7a56209f17a5fbf6717935c7fffad2b1386e50c901ebe" already present on machine kubelet logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:47 +0000 UTC Normal Pod simple-prod-query-755b7d8989-nr9m7.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:47 +0000 UTC Normal Pod simple-prod-query-755b7d8989-nr9m7.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:47 +0000 UTC Normal Pod simple-prod-query-755b7d8989-nr9m7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:47 +0000 UTC Normal ReplicaSet.apps simple-prod-query-755b7d8989 SuccessfulCreate Created pod: simple-prod-query-755b7d8989-nr9m7 replicaset-controller logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:47 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-755b7d8989 to 1 deployment-controller logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:48 +0000 UTC Normal Pod simple-prod-query-755b7d8989-nr9m7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:48 +0000 UTC Normal Pod simple-prod-query-755b7d8989-nr9m7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:48 +0000 UTC Normal Pod simple-prod-query-755b7d8989-nr9m7.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4a07026026451434dc1d4bce30df6e73c9912071b6939b75e35c1526db2ca9e0" already present on machine kubelet logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:48 +0000 UTC Normal Pod simple-prod-query-755b7d8989-nr9m7.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 16:18:52 | collector-autoscale | 2024-06-17 16:18:48 +0000 UTC Normal Pod simple-prod-query-755b7d8989-nr9m7.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 16:18:52 | collector-autoscale | Deleting namespace: kuttl-test-handy-oyster === CONT kuttl/harness/collector-otlp-allinone-grpc logger.go:42: 16:18:58 | collector-otlp-allinone-grpc | Creating namespace: kuttl-test-tidy-ewe logger.go:42: 16:18:58 | collector-otlp-allinone-grpc/0-install | starting test step 0-install logger.go:42: 16:18:58 | collector-otlp-allinone-grpc/0-install | Jaeger:kuttl-test-tidy-ewe/my-jaeger created logger.go:42: 16:19:05 | collector-otlp-allinone-grpc/0-install | test step completed 0-install logger.go:42: 16:19:05 | collector-otlp-allinone-grpc/1-smoke-test | starting test step 1-smoke-test logger.go:42: 16:19:05 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 16:19:06 | collector-otlp-allinone-grpc/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 16:19:13 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 16:19:13 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 16:19:13 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/report-span created logger.go:42: 16:19:13 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/check-span created logger.go:42: 16:19:34 | collector-otlp-allinone-grpc/1-smoke-test | test step completed 1-smoke-test logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | collector-otlp-allinone-grpc events from ns kuttl-test-tidy-ewe: logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:02 +0000 UTC Normal Pod my-jaeger-547444c7c4-b5dzs Binding Scheduled Successfully assigned kuttl-test-tidy-ewe/my-jaeger-547444c7c4-b5dzs to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:02 +0000 UTC Normal ReplicaSet.apps my-jaeger-547444c7c4 SuccessfulCreate Created pod: my-jaeger-547444c7c4-b5dzs replicaset-controller logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:02 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-547444c7c4 to 1 deployment-controller logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:03 +0000 UTC Normal Pod my-jaeger-547444c7c4-b5dzs AddedInterface Add eth0 [10.129.2.69/23] from ovn-kubernetes multus logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:03 +0000 UTC Normal Pod my-jaeger-547444c7c4-b5dzs.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:03 +0000 UTC Normal Pod my-jaeger-547444c7c4-b5dzs.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:03 +0000 UTC Normal Pod my-jaeger-547444c7c4-b5dzs.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:03 +0000 UTC Normal Pod my-jaeger-547444c7c4-b5dzs.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:03 +0000 UTC Normal Pod my-jaeger-547444c7c4-b5dzs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:03 +0000 UTC Normal Pod my-jaeger-547444c7c4-b5dzs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:07 +0000 UTC Normal Pod my-jaeger-547444c7c4-b5dzs.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:07 +0000 UTC Normal Pod my-jaeger-547444c7c4-b5dzs.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:07 +0000 UTC Normal ReplicaSet.apps my-jaeger-547444c7c4 SuccessfulDelete Deleted pod: my-jaeger-547444c7c4-b5dzs replicaset-controller logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:07 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-547444c7c4 to 0 from 1 deployment-controller logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:08 +0000 UTC Normal Pod my-jaeger-59b44ddf89-w7c67 Binding Scheduled Successfully assigned kuttl-test-tidy-ewe/my-jaeger-59b44ddf89-w7c67 to ip-10-0-90-98.us-east-2.compute.internal default-scheduler logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:08 +0000 UTC Normal Pod my-jaeger-59b44ddf89-w7c67 AddedInterface Add eth0 [10.129.2.70/23] from ovn-kubernetes multus logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:08 +0000 UTC Normal ReplicaSet.apps my-jaeger-59b44ddf89 SuccessfulCreate Created pod: my-jaeger-59b44ddf89-w7c67 replicaset-controller logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:08 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-59b44ddf89 to 1 deployment-controller logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:09 +0000 UTC Normal Pod my-jaeger-59b44ddf89-w7c67.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e625cd7ba4aa7d475dfb5264480f79daa7ee071a23b1fc80675d9867443a7c18" already present on machine kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:09 +0000 UTC Normal Pod my-jaeger-59b44ddf89-w7c67.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:09 +0000 UTC Normal Pod my-jaeger-59b44ddf89-w7c67.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:09 +0000 UTC Normal Pod my-jaeger-59b44ddf89-w7c67.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:e52f88555004b575104691fe96f758ee5e6ec832b99107a5a18fc8379d29afda" already present on machine kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:09 +0000 UTC Normal Pod my-jaeger-59b44ddf89-w7c67.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:09 +0000 UTC Normal Pod my-jaeger-59b44ddf89-w7c67.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:13 +0000 UTC Normal Pod check-span-vkbh2 Binding Scheduled Successfully assigned kuttl-test-tidy-ewe/check-span-vkbh2 to ip-10-0-80-215.us-east-2.compute.internal default-scheduler logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:13 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-vkbh2 job-controller logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:13 +0000 UTC Normal Pod report-span-ph25k Binding Scheduled Successfully assigned kuttl-test-tidy-ewe/report-span-ph25k to ip-10-0-38-95.us-east-2.compute.internal default-scheduler logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:13 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-ph25k job-controller logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:14 +0000 UTC Normal Pod check-span-vkbh2 AddedInterface Add eth0 [10.128.2.61/23] from ovn-kubernetes multus logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:14 +0000 UTC Normal Pod check-span-vkbh2.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:14 +0000 UTC Normal Pod check-span-vkbh2.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 340ms (340ms including waiting) kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:14 +0000 UTC Normal Pod check-span-vkbh2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:14 +0000 UTC Normal Pod check-span-vkbh2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:14 +0000 UTC Normal Pod report-span-ph25k AddedInterface Add eth0 [10.131.0.90/23] from ovn-kubernetes multus logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:14 +0000 UTC Normal Pod report-span-ph25k.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:14 +0000 UTC Normal Pod report-span-ph25k.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 349ms (349ms including waiting) kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:14 +0000 UTC Normal Pod report-span-ph25k.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:14 +0000 UTC Normal Pod report-span-ph25k.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | 2024-06-17 16:19:33 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 16:19:34 | collector-otlp-allinone-grpc | Deleting namespace: kuttl-test-tidy-ewe === CONT kuttl/harness/cassandra-spark logger.go:42: 16:19:47 | cassandra-spark | Ignoring 01-assert.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 16:19:47 | cassandra-spark | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 16:19:47 | cassandra-spark | Creating namespace: kuttl-test-promoted-goat logger.go:42: 16:19:47 | cassandra-spark | cassandra-spark events from ns kuttl-test-promoted-goat: logger.go:42: 16:19:47 | cassandra-spark | Deleting namespace: kuttl-test-promoted-goat === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (337.19s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.13s) --- PASS: kuttl/harness/collector-otlp-allinone-http (34.12s) --- PASS: kuttl/harness/set-custom-img (48.34s) --- PASS: kuttl/harness/collector-otlp-production-http (70.26s) --- PASS: kuttl/harness/collector-otlp-production-grpc (78.29s) --- PASS: kuttl/harness/collector-autoscale (45.28s) --- PASS: kuttl/harness/collector-otlp-allinone-grpc (48.28s) --- PASS: kuttl/harness/cassandra-spark (6.30s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name miscellaneous --report --output /logs/artifacts/miscellaneous.xml ./artifacts/kuttl-report.xml time="2024-06-17T16:19:53Z" level=debug msg="Setting a new name for the test suites" time="2024-06-17T16:19:53Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-06-17T16:19:53Z" level=debug msg="normalizing test case names" time="2024-06-17T16:19:53Z" level=debug msg="miscellaneous/artifacts -> miscellaneous_artifacts" time="2024-06-17T16:19:53Z" level=debug msg="miscellaneous/collector-otlp-allinone-http -> miscellaneous_collector_otlp_allinone_http" time="2024-06-17T16:19:53Z" level=debug msg="miscellaneous/set-custom-img -> miscellaneous_set_custom_img" time="2024-06-17T16:19:53Z" level=debug msg="miscellaneous/collector-otlp-production-http -> miscellaneous_collector_otlp_production_http" time="2024-06-17T16:19:53Z" level=debug msg="miscellaneous/collector-otlp-production-grpc -> miscellaneous_collector_otlp_production_grpc" time="2024-06-17T16:19:53Z" level=debug msg="miscellaneous/collector-autoscale -> miscellaneous_collector_autoscale" time="2024-06-17T16:19:53Z" level=debug msg="miscellaneous/collector-otlp-allinone-grpc -> miscellaneous_collector_otlp_allinone_grpc" time="2024-06-17T16:19:53Z" level=debug msg="miscellaneous/cassandra-spark -> miscellaneous_cassandra_spark" +----------------------------------------------+--------+ | NAME | RESULT | +----------------------------------------------+--------+ | miscellaneous_artifacts | passed | | miscellaneous_collector_otlp_allinone_http | passed | | miscellaneous_set_custom_img | passed | | miscellaneous_collector_otlp_production_http | passed | | miscellaneous_collector_otlp_production_grpc | passed | | miscellaneous_collector_autoscale | passed | | miscellaneous_collector_otlp_allinone_grpc | passed | | miscellaneous_cassandra_spark | passed | +----------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests'