% Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 3831 100 3831 0 0 29173 0 --:--:-- --:--:-- --:--:-- 29244 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 9315 100 9315 0 0 77695 0 --:--:-- --:--:-- --:--:-- 78277 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 2607 100 2607 0 0 29591 0 --:--:-- --:--:-- --:--:-- 29625 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 1953 100 1953 0 0 21892 0 --:--:-- --:--:-- --:--:-- 21700 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 350 100 350 0 0 3748 0 --:--:-- --:--:-- --:--:-- 3763 Installing kuttl Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/kubectl-kuttl https://github.com/kudobuilder/kuttl/releases/download/v0.15.0/kubectl-kuttl_0.15.0_linux_x86_64 KUBECONFIG file is: /tmp/kubeconfig-2507913876 for suite in elasticsearch examples generate upgrade sidecar streaming ui miscellaneous; do \ make run-e2e-tests-$suite ; \ done make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh elasticsearch false true + '[' 3 -ne 3 ']' + test_suite_name=elasticsearch + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/elasticsearch.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-elasticsearch make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true \ KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ ./tests/e2e/elasticsearch/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-02-29-062601 True False 11m Cluster version is 4.16.0-0.nightly-2024-02-29-062601' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-02-29-062601 True False 11m Cluster version is 4.16.0-0.nightly-2024-02-29-062601' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/elasticsearch/render.sh ++ export SUITE_DIR=./tests/e2e/elasticsearch ++ SUITE_DIR=./tests/e2e/elasticsearch ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/elasticsearch ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + start_test es-from-aio-to-production + '[' 1 -ne 1 ']' + test_name=es-from-aio-to-production + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-from-aio-to-production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-from-aio-to-production\e[0m' Rendering files for test es-from-aio-to-production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-from-aio-to-production + cd es-from-aio-to-production + jaeger_name=my-jaeger + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 03 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=03 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./03-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch.redundancyPolicy="ZeroRedundancy"' ./03-install.yaml + render_smoke_test my-jaeger true 04 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=04 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./04-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test es-increasing-replicas + '[' 1 -ne 1 ']' + test_name=es-increasing-replicas + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-increasing-replicas' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-increasing-replicas\e[0m' Rendering files for test es-increasing-replicas + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-from-aio-to-production + '[' es-from-aio-to-production '!=' _build ']' + cd .. + mkdir -p es-increasing-replicas + cd es-increasing-replicas + jaeger_name=simple-prod + '[' true = true ']' + jaeger_deployment_mode=production_autoprovisioned + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.replicas=2 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.query.replicas=2 ./02-install.yaml + cp ./01-assert.yaml ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.replicas=2 ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .status.readyReplicas=2 ./02-assert.yaml + render_smoke_test simple-prod true 03 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=03 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./03-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + cp ./02-install.yaml ./04-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.elasticsearch.nodeCount=2 ./04-install.yaml + /tmp/jaeger-tests/bin/gomplate -f ./openshift-check-es-nodes.yaml.template -o ./05-check-es-nodes.yaml + '[' true = true ']' + skip_test es-index-cleaner-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-increasing-replicas + '[' es-increasing-replicas '!=' _build ']' + cd .. + rm -rf es-index-cleaner-upstream + warning 'es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_index_cleaner -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-index-cleaner-autoprov + '[' 1 -ne 1 ']' + test_name=es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-index-cleaner-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-index-cleaner-autoprov\e[0m' Rendering files for test es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-index-cleaner-autoprov + cd es-index-cleaner-autoprov + jaeger_name=test-es-index-cleaner-with-prefix + cronjob_name=test-es-index-cleaner-with-prefix-es-index-cleaner + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + cp ../../es-index-cleaner-upstream/04-assert.yaml ../../es-index-cleaner-upstream/README.md . + render_install_jaeger test-es-index-cleaner-with-prefix production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options.es.index-prefix=""' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.enabled=false ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.numberOfDays=0 ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.esIndexCleaner.schedule="*/1 * * * *"' ./01-install.yaml + render_report_spans test-es-index-cleaner-with-prefix true 5 00 true 02 + '[' 6 -ne 6 ']' + jaeger=test-es-index-cleaner-with-prefix + is_secured=true + number_of_spans=5 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + export JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=5 + DAYS=5 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + sed 's~enabled: false~enabled: true~gi' ./01-install.yaml + CRONJOB_NAME=test-es-index-cleaner-with-prefix-es-index-cleaner + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./04-wait-es-index-cleaner.yaml + /tmp/jaeger-tests/bin/gomplate -f ./01-install.yaml -o ./05-install.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 00 06 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=00 + test_step=06 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=test-es-index-cleaner-with-prefix-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=test-es-index-cleaner-with-prefix-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./06-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./06-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.3 ++ version_ge 5.8.3 5.4 +++ echo 5.8.3 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.3 == 5.8.3 + '[' -n '' ']' + skip_test es-index-cleaner-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-index-cleaner-autoprov + '[' es-index-cleaner-autoprov '!=' _build ']' + cd .. + rm -rf es-index-cleaner-managed + warning 'es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + start_test es-multiinstance + '[' 1 -ne 1 ']' + test_name=es-multiinstance + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-multiinstance' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-multiinstance\e[0m' Rendering files for test es-multiinstance + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-multiinstance + cd es-multiinstance + jaeger_name=instance-1 + render_install_jaeger instance-1 production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=instance-1 + JAEGER_NAME=instance-1 + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f ./03-create-second-instance.yaml.template -o 03-create-second-instance.yaml + '[' true = true ']' + skip_test es-rollover-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-rollover-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-multiinstance + '[' es-multiinstance '!=' _build ']' + cd .. + rm -rf es-rollover-upstream + warning 'es-rollover-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_rollover -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-rollover-autoprov + '[' 1 -ne 1 ']' + test_name=es-rollover-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-rollover-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-rollover-autoprov\e[0m' Rendering files for test es-rollover-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-rollover-autoprov + cd es-rollover-autoprov + cp ../../es-rollover-upstream/05-assert.yaml ../../es-rollover-upstream/05-install.yaml ../../es-rollover-upstream/README.md . + jaeger_name=my-jaeger + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_report_spans my-jaeger true 2 00 true 02 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 00 03 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=00 + test_step=03 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./03-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./03-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 01 04 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=01 + test_step=04 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=01 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./04-check-indices.yaml + JOB_NUMBER=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./04-assert.yaml + render_report_spans my-jaeger true 2 02 true 06 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=02 + ensure_reported_spans=true + test_step=06 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=02 + JOB_NUMBER=02 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./06-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./06-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 02 07 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=02 + test_step=07 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=02 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./07-check-indices.yaml + JOB_NUMBER=02 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./07-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' 03 08 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + job_number=03 + test_step=08 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=03 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./08-check-indices.yaml + JOB_NUMBER=03 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./08-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' 04 09 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + job_number=04 + test_step=09 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=04 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./09-check-indices.yaml + JOB_NUMBER=04 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./09-assert.yaml + render_report_spans my-jaeger true 2 03 true 10 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=03 + ensure_reported_spans=true + test_step=10 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=03 + JOB_NUMBER=03 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./10-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./10-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + CRONJOB_NAME=my-jaeger-es-rollover + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./11-wait-rollover.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-000002'\'',' 05 11 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-000002'\'',' + job_number=05 + test_step=11 + escape_command ''\''--name'\'', '\''jaeger-span-000002'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-000002'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-000002'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-000002'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=05 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./11-check-indices.yaml + JOB_NUMBER=05 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./11-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' 06 12 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + job_number=06 + test_step=12 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=06 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./12-check-indices.yaml + JOB_NUMBER=06 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./12-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.3 ++ version_ge 5.8.3 5.4 +++ echo 5.8.3 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.3 == 5.8.3 + '[' -n '' ']' + skip_test es-rollover-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-rollover-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-rollover-autoprov + '[' es-rollover-autoprov '!=' _build ']' + cd .. + rm -rf es-rollover-managed + warning 'es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + skip_test es-spark-dependencies 'This test is not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=es-spark-dependencies + message='This test is not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + rm -rf es-spark-dependencies + warning 'es-spark-dependencies: This test is not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-spark-dependencies: This test is not supported in OpenShift\e[0m' WAR: es-spark-dependencies: This test is not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running elasticsearch E2E tests' Running elasticsearch E2E tests + cd tests/e2e/elasticsearch/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2507913876 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/es-from-aio-to-production === PAUSE kuttl/harness/es-from-aio-to-production === RUN kuttl/harness/es-increasing-replicas === PAUSE kuttl/harness/es-increasing-replicas === RUN kuttl/harness/es-index-cleaner-autoprov === PAUSE kuttl/harness/es-index-cleaner-autoprov === RUN kuttl/harness/es-multiinstance === PAUSE kuttl/harness/es-multiinstance === RUN kuttl/harness/es-rollover-autoprov === PAUSE kuttl/harness/es-rollover-autoprov === RUN kuttl/harness/es-simple-prod === PAUSE kuttl/harness/es-simple-prod === CONT kuttl/harness/artifacts logger.go:42: 06:59:34 | artifacts | Creating namespace: kuttl-test-amazing-wolf logger.go:42: 06:59:34 | artifacts | artifacts events from ns kuttl-test-amazing-wolf: logger.go:42: 06:59:34 | artifacts | Deleting namespace: kuttl-test-amazing-wolf === CONT kuttl/harness/es-multiinstance logger.go:42: 06:59:41 | es-multiinstance | Ignoring 03-create-second-instance.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 06:59:41 | es-multiinstance | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 06:59:41 | es-multiinstance | Creating namespace: kuttl-test-awaited-locust logger.go:42: 06:59:41 | es-multiinstance/0-clear-namespace | starting test step 0-clear-namespace logger.go:42: 06:59:41 | es-multiinstance/0-clear-namespace | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --ignore-not-found=true] logger.go:42: 06:59:41 | es-multiinstance/0-clear-namespace | test step completed 0-clear-namespace logger.go:42: 06:59:41 | es-multiinstance/1-install | starting test step 1-install logger.go:42: 06:59:41 | es-multiinstance/1-install | Jaeger:kuttl-test-awaited-locust/instance-1 created logger.go:42: 07:00:34 | es-multiinstance/1-install | test step completed 1-install logger.go:42: 07:00:34 | es-multiinstance/2-create-namespace | starting test step 2-create-namespace logger.go:42: 07:00:34 | es-multiinstance/2-create-namespace | running command: [sh -c kubectl create namespace jaeger-e2e-multiinstance-test] logger.go:42: 07:00:34 | es-multiinstance/2-create-namespace | namespace/jaeger-e2e-multiinstance-test created logger.go:42: 07:00:34 | es-multiinstance/2-create-namespace | test step completed 2-create-namespace logger.go:42: 07:00:34 | es-multiinstance/3-create-second-instance | starting test step 3-create-second-instance logger.go:42: 07:00:34 | es-multiinstance/3-create-second-instance | running command: [sh -c kubectl apply -f ./01-install.yaml -n jaeger-e2e-multiinstance-test] logger.go:42: 07:00:38 | es-multiinstance/3-create-second-instance | jaeger.jaegertracing.io/instance-1 created logger.go:42: 07:00:38 | es-multiinstance/3-create-second-instance | running command: [sh -c /tmp/jaeger-tests/bin/kubectl-kuttl assert ./01-assert.yaml -n jaeger-e2e-multiinstance-test --timeout 1000] logger.go:42: 07:01:32 | es-multiinstance/3-create-second-instance | assert is valid logger.go:42: 07:01:32 | es-multiinstance/3-create-second-instance | test step completed 3-create-second-instance logger.go:42: 07:01:32 | es-multiinstance/4-check-secrets | starting test step 4-check-secrets logger.go:42: 07:01:32 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n $NAMESPACE > secret1] logger.go:42: 07:01:32 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n jaeger-e2e-multiinstance-test > secret2] logger.go:42: 07:01:32 | es-multiinstance/4-check-secrets | running command: [sh -c cmp --silent secret1 secret2 || exit 0] logger.go:42: 07:01:32 | es-multiinstance/4-check-secrets | test step completed 4-check-secrets logger.go:42: 07:01:32 | es-multiinstance/5-delete | starting test step 5-delete logger.go:42: 07:01:32 | es-multiinstance/5-delete | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --wait=false] logger.go:42: 07:01:32 | es-multiinstance/5-delete | namespace "jaeger-e2e-multiinstance-test" deleted logger.go:42: 07:01:32 | es-multiinstance/5-delete | test step completed 5-delete logger.go:42: 07:01:32 | es-multiinstance | es-multiinstance events from ns kuttl-test-awaited-locust: logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 06:59:48 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestawaitedlocustinstance1-1-565d64b956 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestawaitedlocustinstance1-1-565d64h8n9n replicaset-controller logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 06:59:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedlocustinstance1-1-565d64h8n9n Binding Scheduled Successfully assigned kuttl-test-awaited-locust/elasticsearch-cdm-kuttltestawaitedlocustinstance1-1-565d64h8n9n to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 06:59:48 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestawaitedlocustinstance1-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestawaitedlocustinstance1-1-565d64b956 to 1 deployment-controller logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 06:59:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedlocustinstance1-1-565d64h8n9n AddedInterface Add eth0 [10.128.2.17/23] from ovn-kubernetes multus logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 06:59:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedlocustinstance1-1-565d64h8n9n.spec.containers{elasticsearch} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 06:59:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedlocustinstance1-1-565d64h8n9n.spec.containers{elasticsearch} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" in 6.675s (6.675s including waiting) kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 06:59:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedlocustinstance1-1-565d64h8n9n.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 06:59:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedlocustinstance1-1-565d64h8n9n.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 06:59:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedlocustinstance1-1-565d64h8n9n.spec.containers{proxy} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 06:59:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedlocustinstance1-1-565d64h8n9n.spec.containers{proxy} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" in 3.567s (3.567s including waiting) kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 06:59:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedlocustinstance1-1-565d64h8n9n.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 06:59:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedlocustinstance1-1-565d64h8n9n.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:09 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestawaitedlocustinstance1-1-565d64h8n9n.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:25 +0000 UTC Normal Pod instance-1-collector-84b889f6b5-l9jwf Binding Scheduled Successfully assigned kuttl-test-awaited-locust/instance-1-collector-84b889f6b5-l9jwf to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:25 +0000 UTC Normal ReplicaSet.apps instance-1-collector-84b889f6b5 SuccessfulCreate Created pod: instance-1-collector-84b889f6b5-l9jwf replicaset-controller logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:25 +0000 UTC Normal Deployment.apps instance-1-collector ScalingReplicaSet Scaled up replica set instance-1-collector-84b889f6b5 to 1 deployment-controller logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:25 +0000 UTC Normal Pod instance-1-query-544c599964-dg5r9 Binding Scheduled Successfully assigned kuttl-test-awaited-locust/instance-1-query-544c599964-dg5r9 to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:25 +0000 UTC Normal ReplicaSet.apps instance-1-query-544c599964 SuccessfulCreate Created pod: instance-1-query-544c599964-dg5r9 replicaset-controller logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:25 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-544c599964 to 1 deployment-controller logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:26 +0000 UTC Normal Pod instance-1-collector-84b889f6b5-l9jwf AddedInterface Add eth0 [10.129.2.23/23] from ovn-kubernetes multus logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:26 +0000 UTC Normal Pod instance-1-collector-84b889f6b5-l9jwf.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:26 +0000 UTC Normal Pod instance-1-query-544c599964-dg5r9 AddedInterface Add eth0 [10.131.0.21/23] from ovn-kubernetes multus logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:26 +0000 UTC Normal Pod instance-1-query-544c599964-dg5r9.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:27 +0000 UTC Normal Pod instance-1-collector-84b889f6b5-l9jwf.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" in 1.76s (1.76s including waiting) kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:28 +0000 UTC Normal Pod instance-1-collector-84b889f6b5-l9jwf.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:28 +0000 UTC Normal Pod instance-1-collector-84b889f6b5-l9jwf.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:30 +0000 UTC Normal Pod instance-1-query-544c599964-dg5r9.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" in 4.12s (4.12s including waiting) kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:30 +0000 UTC Normal Pod instance-1-query-544c599964-dg5r9.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:30 +0000 UTC Normal Pod instance-1-query-544c599964-dg5r9.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:30 +0000 UTC Normal Pod instance-1-query-544c599964-dg5r9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:30 +0000 UTC Normal Pod instance-1-query-544c599964-dg5r9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:30 +0000 UTC Normal Pod instance-1-query-544c599964-dg5r9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:30 +0000 UTC Normal Pod instance-1-query-544c599964-dg5r9.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:32 +0000 UTC Normal Pod instance-1-query-544c599964-dg5r9.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" in 1.549s (1.549s including waiting) kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:32 +0000 UTC Normal Pod instance-1-query-544c599964-dg5r9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:32 +0000 UTC Normal Pod instance-1-query-544c599964-dg5r9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:50 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:50 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:00:50 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:01:21 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:01:21 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod instance-1-collector-84b889f6b5-l9jwf horizontal-pod-autoscaler logger.go:42: 07:01:32 | es-multiinstance | 2024-03-04 07:01:21 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:01:32 | es-multiinstance | Deleting namespace: kuttl-test-awaited-locust === CONT kuttl/harness/es-simple-prod logger.go:42: 07:01:39 | es-simple-prod | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:01:39 | es-simple-prod | Creating namespace: kuttl-test-fond-ray logger.go:42: 07:01:39 | es-simple-prod | es-simple-prod events from ns kuttl-test-fond-ray: logger.go:42: 07:01:39 | es-simple-prod | Deleting namespace: kuttl-test-fond-ray === CONT kuttl/harness/es-rollover-autoprov logger.go:42: 07:01:45 | es-rollover-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:01:45 | es-rollover-autoprov | Creating namespace: kuttl-test-saving-sunfish logger.go:42: 07:01:45 | es-rollover-autoprov/1-install | starting test step 1-install logger.go:42: 07:01:45 | es-rollover-autoprov/1-install | Jaeger:kuttl-test-saving-sunfish/my-jaeger created logger.go:42: 07:02:27 | es-rollover-autoprov/1-install | test step completed 1-install logger.go:42: 07:02:27 | es-rollover-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:02:27 | es-rollover-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:02:29 | es-rollover-autoprov/2-report-spans | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:02:37 | es-rollover-autoprov/2-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:02:37 | es-rollover-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:02:38 | es-rollover-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:03:03 | es-rollover-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 07:03:03 | es-rollover-autoprov/3-check-indices | starting test step 3-check-indices logger.go:42: 07:03:03 | es-rollover-autoprov/3-check-indices | Job:kuttl-test-saving-sunfish/00-check-indices created logger.go:42: 07:03:07 | es-rollover-autoprov/3-check-indices | test step completed 3-check-indices logger.go:42: 07:03:07 | es-rollover-autoprov/4-check-indices | starting test step 4-check-indices logger.go:42: 07:03:07 | es-rollover-autoprov/4-check-indices | Job:kuttl-test-saving-sunfish/01-check-indices created logger.go:42: 07:03:11 | es-rollover-autoprov/4-check-indices | test step completed 4-check-indices logger.go:42: 07:03:11 | es-rollover-autoprov/5-install | starting test step 5-install logger.go:42: 07:03:11 | es-rollover-autoprov/5-install | Jaeger:kuttl-test-saving-sunfish/my-jaeger updated logger.go:42: 07:03:21 | es-rollover-autoprov/5-install | test step completed 5-install logger.go:42: 07:03:21 | es-rollover-autoprov/6-report-spans | starting test step 6-report-spans logger.go:42: 07:03:21 | es-rollover-autoprov/6-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:03:29 | es-rollover-autoprov/6-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f JOB_NUMBER=02 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-02-job.yaml] logger.go:42: 07:03:29 | es-rollover-autoprov/6-report-spans | running command: [sh -c kubectl apply -f report-span-02-job.yaml -n $NAMESPACE] logger.go:42: 07:03:29 | es-rollover-autoprov/6-report-spans | job.batch/02-report-span created logger.go:42: 07:03:55 | es-rollover-autoprov/6-report-spans | test step completed 6-report-spans logger.go:42: 07:03:55 | es-rollover-autoprov/7-check-indices | starting test step 7-check-indices logger.go:42: 07:03:55 | es-rollover-autoprov/7-check-indices | Job:kuttl-test-saving-sunfish/02-check-indices created logger.go:42: 07:03:59 | es-rollover-autoprov/7-check-indices | test step completed 7-check-indices logger.go:42: 07:03:59 | es-rollover-autoprov/8-check-indices | starting test step 8-check-indices logger.go:42: 07:03:59 | es-rollover-autoprov/8-check-indices | Job:kuttl-test-saving-sunfish/03-check-indices created logger.go:42: 07:04:03 | es-rollover-autoprov/8-check-indices | test step completed 8-check-indices logger.go:42: 07:04:03 | es-rollover-autoprov/9-check-indices | starting test step 9-check-indices logger.go:42: 07:04:03 | es-rollover-autoprov/9-check-indices | Job:kuttl-test-saving-sunfish/04-check-indices created logger.go:42: 07:04:07 | es-rollover-autoprov/9-check-indices | test step completed 9-check-indices logger.go:42: 07:04:07 | es-rollover-autoprov/10-report-spans | starting test step 10-report-spans logger.go:42: 07:04:07 | es-rollover-autoprov/10-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:04:16 | es-rollover-autoprov/10-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f JOB_NUMBER=03 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-03-job.yaml] logger.go:42: 07:04:17 | es-rollover-autoprov/10-report-spans | running command: [sh -c kubectl apply -f report-span-03-job.yaml -n $NAMESPACE] logger.go:42: 07:04:17 | es-rollover-autoprov/10-report-spans | job.batch/03-report-span created logger.go:42: 07:04:41 | es-rollover-autoprov/10-report-spans | test step completed 10-report-spans logger.go:42: 07:04:41 | es-rollover-autoprov/11-check-indices | starting test step 11-check-indices logger.go:42: 07:04:41 | es-rollover-autoprov/11-check-indices | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob my-jaeger-es-rollover --namespace $NAMESPACE] logger.go:42: 07:04:51 | es-rollover-autoprov/11-check-indices | time="2024-03-04T07:04:51Z" level=debug msg="Checking if the my-jaeger-es-rollover CronJob exists" logger.go:42: 07:04:51 | es-rollover-autoprov/11-check-indices | time="2024-03-04T07:04:51Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 07:04:51 | es-rollover-autoprov/11-check-indices | time="2024-03-04T07:04:51Z" level=info msg="Cronjob my-jaeger-es-rollover found successfully" logger.go:42: 07:04:51 | es-rollover-autoprov/11-check-indices | time="2024-03-04T07:04:51Z" level=debug msg="Waiting for the next scheduled job from my-jaeger-es-rollover cronjob" logger.go:42: 07:04:51 | es-rollover-autoprov/11-check-indices | time="2024-03-04T07:04:51Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:05:01 | es-rollover-autoprov/11-check-indices | time="2024-03-04T07:05:01Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:05:11 | es-rollover-autoprov/11-check-indices | time="2024-03-04T07:05:11Z" level=info msg="Job of owner my-jaeger-es-rollover succeeded after my-jaeger-es-rollover 20.018498618s" logger.go:42: 07:05:11 | es-rollover-autoprov/11-check-indices | Job:kuttl-test-saving-sunfish/05-check-indices created logger.go:42: 07:05:15 | es-rollover-autoprov/11-check-indices | test step completed 11-check-indices logger.go:42: 07:05:15 | es-rollover-autoprov/12-check-indices | starting test step 12-check-indices logger.go:42: 07:05:15 | es-rollover-autoprov/12-check-indices | Job:kuttl-test-saving-sunfish/06-check-indices created logger.go:42: 07:05:19 | es-rollover-autoprov/12-check-indices | test step completed 12-check-indices logger.go:42: 07:05:19 | es-rollover-autoprov | es-rollover-autoprov events from ns kuttl-test-saving-sunfish: logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:01:52 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestsavingsunfishmyjaeger-1-78ffc48846 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestsavingsunfishmyjaeger-1-78ffc48k5gff replicaset-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:01:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsavingsunfishmyjaeger-1-78ffc48k5gff Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/elasticsearch-cdm-kuttltestsavingsunfishmyjaeger-1-78ffc48k5gff to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:01:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsavingsunfishmyjaeger-1-78ffc48k5gff AddedInterface Add eth0 [10.128.2.18/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:01:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsavingsunfishmyjaeger-1-78ffc48k5gff.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:01:52 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestsavingsunfishmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestsavingsunfishmyjaeger-1-78ffc48846 to 1 deployment-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:01:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsavingsunfishmyjaeger-1-78ffc48k5gff.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:01:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsavingsunfishmyjaeger-1-78ffc48k5gff.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:01:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsavingsunfishmyjaeger-1-78ffc48k5gff.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:01:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsavingsunfishmyjaeger-1-78ffc48k5gff.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:01:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsavingsunfishmyjaeger-1-78ffc48k5gff.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:07 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsavingsunfishmyjaeger-1-78ffc48k5gff.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:19 +0000 UTC Normal Pod my-jaeger-collector-7dddfd87f8-mslls Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/my-jaeger-collector-7dddfd87f8-mslls to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:19 +0000 UTC Normal Pod my-jaeger-collector-7dddfd87f8-mslls AddedInterface Add eth0 [10.131.0.24/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:19 +0000 UTC Normal Pod my-jaeger-collector-7dddfd87f8-mslls.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:19 +0000 UTC Normal Pod my-jaeger-collector-7dddfd87f8-mslls.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:19 +0000 UTC Normal Pod my-jaeger-collector-7dddfd87f8-mslls.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:19 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7dddfd87f8 SuccessfulCreate Created pod: my-jaeger-collector-7dddfd87f8-mslls replicaset-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:19 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7dddfd87f8 to 1 deployment-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:19 +0000 UTC Normal Pod my-jaeger-query-55f5fd7df-rj28p Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/my-jaeger-query-55f5fd7df-rj28p to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:19 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-55f5fd7df SuccessfulCreate Created pod: my-jaeger-query-55f5fd7df-rj28p replicaset-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:19 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-55f5fd7df to 1 deployment-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:20 +0000 UTC Normal Pod my-jaeger-query-55f5fd7df-rj28p AddedInterface Add eth0 [10.129.2.25/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:20 +0000 UTC Normal Pod my-jaeger-query-55f5fd7df-rj28p.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:23 +0000 UTC Normal Pod my-jaeger-query-55f5fd7df-rj28p.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" in 2.92s (2.92s including waiting) kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:23 +0000 UTC Normal Pod my-jaeger-query-55f5fd7df-rj28p.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:23 +0000 UTC Normal Pod my-jaeger-query-55f5fd7df-rj28p.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:23 +0000 UTC Normal Pod my-jaeger-query-55f5fd7df-rj28p.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:23 +0000 UTC Normal Pod my-jaeger-query-55f5fd7df-rj28p.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:23 +0000 UTC Normal Pod my-jaeger-query-55f5fd7df-rj28p.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:23 +0000 UTC Normal Pod my-jaeger-query-55f5fd7df-rj28p.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:26 +0000 UTC Normal Pod my-jaeger-query-55f5fd7df-rj28p.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" in 2.786s (2.786s including waiting) kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:26 +0000 UTC Normal Pod my-jaeger-query-55f5fd7df-rj28p.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:26 +0000 UTC Normal Pod my-jaeger-query-55f5fd7df-rj28p.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:32 +0000 UTC Normal Pod my-jaeger-query-55f5fd7df-rj28p.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:32 +0000 UTC Normal Pod my-jaeger-query-55f5fd7df-rj28p.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:32 +0000 UTC Normal Pod my-jaeger-query-55f5fd7df-rj28p.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:32 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-55f5fd7df SuccessfulDelete Deleted pod: my-jaeger-query-55f5fd7df-rj28p replicaset-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:32 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-55f5fd7df to 0 from 1 deployment-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:33 +0000 UTC Normal Pod my-jaeger-query-c788d7998-nksnb Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/my-jaeger-query-c788d7998-nksnb to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:33 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-c788d7998 SuccessfulCreate Created pod: my-jaeger-query-c788d7998-nksnb replicaset-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:33 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-c788d7998 to 1 deployment-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:34 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:34 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:34 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:34 +0000 UTC Normal Pod my-jaeger-query-c788d7998-nksnb AddedInterface Add eth0 [10.129.2.26/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:34 +0000 UTC Normal Pod my-jaeger-query-c788d7998-nksnb.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:34 +0000 UTC Normal Pod my-jaeger-query-c788d7998-nksnb.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:34 +0000 UTC Normal Pod my-jaeger-query-c788d7998-nksnb.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:34 +0000 UTC Normal Pod my-jaeger-query-c788d7998-nksnb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:34 +0000 UTC Normal Pod my-jaeger-query-c788d7998-nksnb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:34 +0000 UTC Normal Pod my-jaeger-query-c788d7998-nksnb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:34 +0000 UTC Normal Pod my-jaeger-query-c788d7998-nksnb.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:34 +0000 UTC Normal Pod my-jaeger-query-c788d7998-nksnb.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:34 +0000 UTC Normal Pod my-jaeger-query-c788d7998-nksnb.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:38 +0000 UTC Normal Pod 00-report-span-fm9kb Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/00-report-span-fm9kb to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:38 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-fm9kb job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:39 +0000 UTC Normal Pod 00-report-span-fm9kb AddedInterface Add eth0 [10.131.0.25/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:39 +0000 UTC Normal Pod 00-report-span-fm9kb.spec.containers{asserts-container} Pulling Pulling image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:40 +0000 UTC Normal Pod 00-report-span-fm9kb.spec.containers{asserts-container} Pulled Successfully pulled image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" in 1.322s (1.322s including waiting) kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:40 +0000 UTC Normal Pod 00-report-span-fm9kb.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:02:40 +0000 UTC Normal Pod 00-report-span-fm9kb.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:03 +0000 UTC Normal Pod 00-check-indices-mxqbj Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/00-check-indices-mxqbj to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:03 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-mxqbj job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:03 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:04 +0000 UTC Normal Pod 00-check-indices-mxqbj AddedInterface Add eth0 [10.131.0.26/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:04 +0000 UTC Normal Pod 00-check-indices-mxqbj.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:04 +0000 UTC Normal Pod 00-check-indices-mxqbj.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:04 +0000 UTC Normal Pod 00-check-indices-mxqbj.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:07 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:07 +0000 UTC Normal Pod 01-check-indices-zzkzd Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/01-check-indices-zzkzd to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:07 +0000 UTC Normal Job.batch 01-check-indices SuccessfulCreate Created pod: 01-check-indices-zzkzd job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:08 +0000 UTC Normal Pod 01-check-indices-zzkzd AddedInterface Add eth0 [10.131.0.27/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:08 +0000 UTC Normal Pod 01-check-indices-zzkzd.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:08 +0000 UTC Normal Pod 01-check-indices-zzkzd.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:08 +0000 UTC Normal Pod 01-check-indices-zzkzd.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:11 +0000 UTC Normal Job.batch 01-check-indices Completed Job completed job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:12 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-8m5lc Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/my-jaeger-es-rollover-create-mapping-8m5lc to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:12 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping SuccessfulCreate Created pod: my-jaeger-es-rollover-create-mapping-8m5lc job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:13 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-8m5lc AddedInterface Add eth0 [10.131.0.28/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:13 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-8m5lc.spec.containers{my-jaeger-es-rollover-create-mapping} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:7e3e0cb8a24b89e2c39c7bfc837f1a8901f0835f204b8b213c7de7c1faaeb171" kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:17 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-8m5lc.spec.containers{my-jaeger-es-rollover-create-mapping} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:7e3e0cb8a24b89e2c39c7bfc837f1a8901f0835f204b8b213c7de7c1faaeb171" in 4.461s (4.461s including waiting) kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:17 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-8m5lc.spec.containers{my-jaeger-es-rollover-create-mapping} Created Created container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:17 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-8m5lc.spec.containers{my-jaeger-es-rollover-create-mapping} Started Started container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:21 +0000 UTC Normal Pod my-jaeger-collector-7dddfd87f8-mslls.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:21 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7dddfd87f8 SuccessfulDelete Deleted pod: my-jaeger-collector-7dddfd87f8-mslls replicaset-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:21 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-7dddfd87f8 to 0 from 1 deployment-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:21 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping Completed Job completed job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:21 +0000 UTC Normal Pod my-jaeger-query-c788d7998-nksnb.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:21 +0000 UTC Normal Pod my-jaeger-query-c788d7998-nksnb.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:21 +0000 UTC Normal Pod my-jaeger-query-c788d7998-nksnb.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:21 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-c788d7998 SuccessfulDelete Deleted pod: my-jaeger-query-c788d7998-nksnb replicaset-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:21 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-c788d7998 to 0 from 1 deployment-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:22 +0000 UTC Normal Pod my-jaeger-collector-d7b75b7c9-82d5g Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/my-jaeger-collector-d7b75b7c9-82d5g to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:22 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-d7b75b7c9 SuccessfulCreate Created pod: my-jaeger-collector-d7b75b7c9-82d5g replicaset-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:22 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-d7b75b7c9 to 1 deployment-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:22 +0000 UTC Normal Pod my-jaeger-query-69f684b9c-d6bn7 Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/my-jaeger-query-69f684b9c-d6bn7 to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:22 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-69f684b9c SuccessfulCreate Created pod: my-jaeger-query-69f684b9c-d6bn7 replicaset-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:22 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-69f684b9c to 1 deployment-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:23 +0000 UTC Normal Pod my-jaeger-collector-d7b75b7c9-82d5g AddedInterface Add eth0 [10.129.2.27/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:23 +0000 UTC Normal Pod my-jaeger-collector-d7b75b7c9-82d5g.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:23 +0000 UTC Normal Pod my-jaeger-collector-d7b75b7c9-82d5g.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:23 +0000 UTC Normal Pod my-jaeger-collector-d7b75b7c9-82d5g.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:23 +0000 UTC Normal Pod my-jaeger-query-69f684b9c-d6bn7 AddedInterface Add eth0 [10.131.0.29/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:23 +0000 UTC Normal Pod my-jaeger-query-69f684b9c-d6bn7.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:23 +0000 UTC Normal Pod my-jaeger-query-69f684b9c-d6bn7.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:23 +0000 UTC Normal Pod my-jaeger-query-69f684b9c-d6bn7.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:23 +0000 UTC Normal Pod my-jaeger-query-69f684b9c-d6bn7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:23 +0000 UTC Normal Pod my-jaeger-query-69f684b9c-d6bn7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:23 +0000 UTC Normal Pod my-jaeger-query-69f684b9c-d6bn7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:23 +0000 UTC Normal Pod my-jaeger-query-69f684b9c-d6bn7.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:23 +0000 UTC Normal Pod my-jaeger-query-69f684b9c-d6bn7.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:23 +0000 UTC Normal Pod my-jaeger-query-69f684b9c-d6bn7.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:29 +0000 UTC Normal Pod 02-report-span-zdkhs Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/02-report-span-zdkhs to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:29 +0000 UTC Normal Job.batch 02-report-span SuccessfulCreate Created pod: 02-report-span-zdkhs job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:30 +0000 UTC Normal Pod 02-report-span-zdkhs AddedInterface Add eth0 [10.129.2.28/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:30 +0000 UTC Normal Pod 02-report-span-zdkhs.spec.containers{asserts-container} Pulling Pulling image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:31 +0000 UTC Normal Pod 02-report-span-zdkhs.spec.containers{asserts-container} Pulled Successfully pulled image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" in 1.206s (1.206s including waiting) kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:31 +0000 UTC Normal Pod 02-report-span-zdkhs.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:31 +0000 UTC Normal Pod 02-report-span-zdkhs.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:54 +0000 UTC Normal Job.batch 02-report-span Completed Job completed job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:55 +0000 UTC Normal Pod 02-check-indices-zwmb2 Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/02-check-indices-zwmb2 to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:55 +0000 UTC Normal Pod 02-check-indices-zwmb2 AddedInterface Add eth0 [10.129.2.29/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:55 +0000 UTC Normal Pod 02-check-indices-zwmb2.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:55 +0000 UTC Normal Pod 02-check-indices-zwmb2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:55 +0000 UTC Normal Pod 02-check-indices-zwmb2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:55 +0000 UTC Normal Job.batch 02-check-indices SuccessfulCreate Created pod: 02-check-indices-zwmb2 job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:58 +0000 UTC Normal Job.batch 02-check-indices Completed Job completed job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:59 +0000 UTC Normal Pod 03-check-indices-hz7gc Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/03-check-indices-hz7gc to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:59 +0000 UTC Normal Pod 03-check-indices-hz7gc AddedInterface Add eth0 [10.129.2.30/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:59 +0000 UTC Normal Pod 03-check-indices-hz7gc.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:59 +0000 UTC Normal Pod 03-check-indices-hz7gc.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:59 +0000 UTC Normal Pod 03-check-indices-hz7gc.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:03:59 +0000 UTC Normal Job.batch 03-check-indices SuccessfulCreate Created pod: 03-check-indices-hz7gc job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28492264-m9v8l Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/my-jaeger-es-lookback-28492264-m9v8l to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28492264-m9v8l AddedInterface Add eth0 [10.129.2.31/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28492264-m9v8l.spec.containers{my-jaeger-es-lookback} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:7e3e0cb8a24b89e2c39c7bfc837f1a8901f0835f204b8b213c7de7c1faaeb171" kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28492264 SuccessfulCreate Created pod: my-jaeger-es-lookback-28492264-m9v8l job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28492264 cronjob-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28492264-59dcw Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/my-jaeger-es-rollover-28492264-59dcw to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28492264-59dcw AddedInterface Add eth0 [10.131.0.30/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28492264-59dcw.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:7e3e0cb8a24b89e2c39c7bfc837f1a8901f0835f204b8b213c7de7c1faaeb171" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28492264-59dcw.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28492264-59dcw.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28492264 SuccessfulCreate Created pod: my-jaeger-es-rollover-28492264-59dcw job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28492264 cronjob-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:03 +0000 UTC Normal Job.batch 03-check-indices Completed Job completed job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:03 +0000 UTC Normal Pod 04-check-indices-76dtv Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/04-check-indices-76dtv to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:03 +0000 UTC Normal Pod 04-check-indices-76dtv AddedInterface Add eth0 [10.129.2.32/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:03 +0000 UTC Normal Pod 04-check-indices-76dtv.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:03 +0000 UTC Normal Pod 04-check-indices-76dtv.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:03 +0000 UTC Normal Job.batch 04-check-indices SuccessfulCreate Created pod: 04-check-indices-76dtv job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28492264 Completed Job completed job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28492264, status: Complete cronjob-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:04 +0000 UTC Normal Pod 04-check-indices-76dtv.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:07 +0000 UTC Normal Job.batch 04-check-indices Completed Job completed job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:15 +0000 UTC Normal Pod my-jaeger-es-lookback-28492264-m9v8l.spec.containers{my-jaeger-es-lookback} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:7e3e0cb8a24b89e2c39c7bfc837f1a8901f0835f204b8b213c7de7c1faaeb171" in 14.972s (14.972s including waiting) kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:15 +0000 UTC Normal Pod my-jaeger-es-lookback-28492264-m9v8l.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:15 +0000 UTC Normal Pod my-jaeger-es-lookback-28492264-m9v8l.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:17 +0000 UTC Normal Pod 03-report-span-mchxr Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/03-report-span-mchxr to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:17 +0000 UTC Normal Job.batch 03-report-span SuccessfulCreate Created pod: 03-report-span-mchxr job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:18 +0000 UTC Normal Pod 03-report-span-mchxr AddedInterface Add eth0 [10.129.2.33/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:18 +0000 UTC Normal Pod 03-report-span-mchxr.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:18 +0000 UTC Normal Pod 03-report-span-mchxr.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:18 +0000 UTC Normal Pod 03-report-span-mchxr.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:18 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28492264 Completed Job completed job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:18 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28492264, status: Complete cronjob-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-d7b75b7c9-82d5g horizontal-pod-autoscaler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:04:41 +0000 UTC Normal Job.batch 03-report-span Completed Job completed job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28492265-k6d2x Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/my-jaeger-es-lookback-28492265-k6d2x to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28492265-k6d2x AddedInterface Add eth0 [10.129.2.34/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28492265-k6d2x.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:7e3e0cb8a24b89e2c39c7bfc837f1a8901f0835f204b8b213c7de7c1faaeb171" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28492265-k6d2x.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28492265-k6d2x.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28492265 SuccessfulCreate Created pod: my-jaeger-es-lookback-28492265-k6d2x job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28492265 cronjob-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28492265-w2kzv Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/my-jaeger-es-rollover-28492265-w2kzv to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28492265-w2kzv AddedInterface Add eth0 [10.129.2.35/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28492265-w2kzv.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:7e3e0cb8a24b89e2c39c7bfc837f1a8901f0835f204b8b213c7de7c1faaeb171" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28492265-w2kzv.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28492265-w2kzv.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28492265 SuccessfulCreate Created pod: my-jaeger-es-rollover-28492265-w2kzv job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28492265 cronjob-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28492265 Completed Job completed job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28492265, status: Complete cronjob-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28492265 Completed Job completed job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28492265, status: Complete cronjob-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:11 +0000 UTC Normal Pod 05-check-indices-hvhvg Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/05-check-indices-hvhvg to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:11 +0000 UTC Normal Pod 05-check-indices-hvhvg AddedInterface Add eth0 [10.129.2.36/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:11 +0000 UTC Normal Job.batch 05-check-indices SuccessfulCreate Created pod: 05-check-indices-hvhvg job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:12 +0000 UTC Normal Pod 05-check-indices-hvhvg.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:12 +0000 UTC Normal Pod 05-check-indices-hvhvg.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:12 +0000 UTC Normal Pod 05-check-indices-hvhvg.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:15 +0000 UTC Normal Job.batch 05-check-indices Completed Job completed job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:15 +0000 UTC Normal Pod 06-check-indices-zcxqf Binding Scheduled Successfully assigned kuttl-test-saving-sunfish/06-check-indices-zcxqf to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:15 +0000 UTC Normal Job.batch 06-check-indices SuccessfulCreate Created pod: 06-check-indices-zcxqf job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:16 +0000 UTC Normal Pod 06-check-indices-zcxqf AddedInterface Add eth0 [10.129.2.37/23] from ovn-kubernetes multus logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:16 +0000 UTC Normal Pod 06-check-indices-zcxqf.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:16 +0000 UTC Normal Pod 06-check-indices-zcxqf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:16 +0000 UTC Normal Pod 06-check-indices-zcxqf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:05:19 | es-rollover-autoprov | 2024-03-04 07:05:19 +0000 UTC Normal Job.batch 06-check-indices Completed Job completed job-controller logger.go:42: 07:05:19 | es-rollover-autoprov | Deleting namespace: kuttl-test-saving-sunfish === CONT kuttl/harness/es-increasing-replicas logger.go:42: 07:05:26 | es-increasing-replicas | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:05:26 | es-increasing-replicas | Ignoring check-es-nodes.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:05:26 | es-increasing-replicas | Ignoring openshift-check-es-nodes.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:05:26 | es-increasing-replicas | Creating namespace: kuttl-test-present-impala logger.go:42: 07:05:26 | es-increasing-replicas/1-install | starting test step 1-install logger.go:42: 07:05:26 | es-increasing-replicas/1-install | Jaeger:kuttl-test-present-impala/simple-prod created logger.go:42: 07:06:03 | es-increasing-replicas/1-install | test step completed 1-install logger.go:42: 07:06:03 | es-increasing-replicas/2-install | starting test step 2-install logger.go:42: 07:06:03 | es-increasing-replicas/2-install | Jaeger:kuttl-test-present-impala/simple-prod updated logger.go:42: 07:06:09 | es-increasing-replicas/2-install | test step completed 2-install logger.go:42: 07:06:09 | es-increasing-replicas/3-smoke-test | starting test step 3-smoke-test logger.go:42: 07:06:09 | es-increasing-replicas/3-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:06:10 | es-increasing-replicas/3-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:06:16 | es-increasing-replicas/3-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:06:17 | es-increasing-replicas/3-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:06:17 | es-increasing-replicas/3-smoke-test | job.batch/report-span created logger.go:42: 07:06:17 | es-increasing-replicas/3-smoke-test | job.batch/check-span created logger.go:42: 07:06:29 | es-increasing-replicas/3-smoke-test | test step completed 3-smoke-test logger.go:42: 07:06:29 | es-increasing-replicas/4-install | starting test step 4-install logger.go:42: 07:06:29 | es-increasing-replicas/4-install | Jaeger:kuttl-test-present-impala/simple-prod updated logger.go:42: 07:06:29 | es-increasing-replicas/4-install | test step completed 4-install logger.go:42: 07:06:29 | es-increasing-replicas/5-check-es-nodes | starting test step 5-check-es-nodes logger.go:42: 07:06:29 | es-increasing-replicas/5-check-es-nodes | running command: [sh -c ./check-es-nodes.sh $NAMESPACE] logger.go:42: 07:06:29 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:06:29 | es-increasing-replicas/5-check-es-nodes | false logger.go:42: 07:06:29 | es-increasing-replicas/5-check-es-nodes | Error: no matches found logger.go:42: 07:06:34 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:06:35 | es-increasing-replicas/5-check-es-nodes | true logger.go:42: 07:06:35 | es-increasing-replicas/5-check-es-nodes | test step completed 5-check-es-nodes logger.go:42: 07:06:35 | es-increasing-replicas | es-increasing-replicas events from ns kuttl-test-present-impala: logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:32 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestpresentimpalasimpleprod-1-5ddcbc4dbd SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestpresentimpalasimpleprod-1-5ddcbrlcb4 replicaset-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-1-5ddcbrlcb4 Binding Scheduled Successfully assigned kuttl-test-present-impala/elasticsearch-cdm-kuttltestpresentimpalasimpleprod-1-5ddcbrlcb4 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:32 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-1-5ddcbrlcb4 FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:32 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestpresentimpalasimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestpresentimpalasimpleprod-1-5ddcbc4dbd to 1 deployment-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-1-5ddcbrlcb4 AddedInterface Add eth0 [10.128.2.19/23] from ovn-kubernetes multus logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-1-5ddcbrlcb4.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-1-5ddcbrlcb4.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-1-5ddcbrlcb4.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-1-5ddcbrlcb4.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-1-5ddcbrlcb4.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-1-5ddcbrlcb4.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:43 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-1-5ddcbrlcb4.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:48 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-1-5ddcbrlcb4.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:59 +0000 UTC Normal Pod simple-prod-collector-7884945cf7-scz8w Binding Scheduled Successfully assigned kuttl-test-present-impala/simple-prod-collector-7884945cf7-scz8w to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:59 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-7884945cf7 SuccessfulCreate Created pod: simple-prod-collector-7884945cf7-scz8w replicaset-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:59 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-7884945cf7 to 1 deployment-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:59 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-28nnx Binding Scheduled Successfully assigned kuttl-test-present-impala/simple-prod-query-6dcbfcdfbb-28nnx to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:59 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6dcbfcdfbb SuccessfulCreate Created pod: simple-prod-query-6dcbfcdfbb-28nnx replicaset-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:05:59 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6dcbfcdfbb to 1 deployment-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:00 +0000 UTC Normal Pod simple-prod-collector-7884945cf7-scz8w AddedInterface Add eth0 [10.131.0.31/23] from ovn-kubernetes multus logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:00 +0000 UTC Normal Pod simple-prod-collector-7884945cf7-scz8w.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:00 +0000 UTC Normal Pod simple-prod-collector-7884945cf7-scz8w.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:00 +0000 UTC Normal Pod simple-prod-collector-7884945cf7-scz8w.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:00 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-28nnx AddedInterface Add eth0 [10.129.2.38/23] from ovn-kubernetes multus logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:00 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-28nnx.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:00 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-28nnx.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:00 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-28nnx.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:00 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-28nnx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:00 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-28nnx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:00 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-28nnx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:00 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-28nnx.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:00 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-28nnx.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:00 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-28nnx.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:06 +0000 UTC Normal Pod simple-prod-collector-7884945cf7-flqzt Binding Scheduled Successfully assigned kuttl-test-present-impala/simple-prod-collector-7884945cf7-flqzt to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:06 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-7884945cf7 SuccessfulCreate Created pod: simple-prod-collector-7884945cf7-flqzt replicaset-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:06 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-7884945cf7 to 2 from 1 deployment-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:06 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-vprws Binding Scheduled Successfully assigned kuttl-test-present-impala/simple-prod-query-6dcbfcdfbb-vprws to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:06 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6dcbfcdfbb SuccessfulCreate Created pod: simple-prod-query-6dcbfcdfbb-vprws replicaset-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:06 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6dcbfcdfbb to 2 from 1 deployment-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:07 +0000 UTC Normal Pod simple-prod-collector-7884945cf7-flqzt AddedInterface Add eth0 [10.129.2.39/23] from ovn-kubernetes multus logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:07 +0000 UTC Normal Pod simple-prod-collector-7884945cf7-flqzt.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:07 +0000 UTC Normal Pod simple-prod-collector-7884945cf7-flqzt.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:07 +0000 UTC Normal Pod simple-prod-collector-7884945cf7-flqzt.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:07 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-vprws AddedInterface Add eth0 [10.131.0.32/23] from ovn-kubernetes multus logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:07 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-vprws.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:07 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-vprws.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:07 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-vprws.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:07 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-vprws.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:07 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-vprws.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:07 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-vprws.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:07 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-vprws.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:07 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-vprws.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:07 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-vprws.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:12 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-28nnx.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:12 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-28nnx.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:12 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-28nnx.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:12 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-vprws.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:12 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-vprws.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:12 +0000 UTC Normal Pod simple-prod-query-6dcbfcdfbb-vprws.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:12 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6dcbfcdfbb SuccessfulDelete Deleted pod: simple-prod-query-6dcbfcdfbb-28nnx replicaset-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:12 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6dcbfcdfbb SuccessfulDelete Deleted pod: simple-prod-query-6dcbfcdfbb-vprws replicaset-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:12 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-6dcbfcdfbb to 0 from 2 deployment-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:13 +0000 UTC Normal Pod simple-prod-query-849456d7cd-fbxv5 Binding Scheduled Successfully assigned kuttl-test-present-impala/simple-prod-query-849456d7cd-fbxv5 to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:13 +0000 UTC Normal Pod simple-prod-query-849456d7cd-fbxv5 AddedInterface Add eth0 [10.131.0.33/23] from ovn-kubernetes multus logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:13 +0000 UTC Normal Pod simple-prod-query-849456d7cd-fbxv5.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:13 +0000 UTC Normal Pod simple-prod-query-849456d7cd-jw52w Binding Scheduled Successfully assigned kuttl-test-present-impala/simple-prod-query-849456d7cd-jw52w to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:13 +0000 UTC Normal Pod simple-prod-query-849456d7cd-jw52w AddedInterface Add eth0 [10.129.2.40/23] from ovn-kubernetes multus logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:13 +0000 UTC Normal ReplicaSet.apps simple-prod-query-849456d7cd SuccessfulCreate Created pod: simple-prod-query-849456d7cd-fbxv5 replicaset-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:13 +0000 UTC Normal ReplicaSet.apps simple-prod-query-849456d7cd SuccessfulCreate Created pod: simple-prod-query-849456d7cd-jw52w replicaset-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:13 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-849456d7cd to 2 deployment-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:14 +0000 UTC Normal Pod simple-prod-query-849456d7cd-fbxv5.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:14 +0000 UTC Normal Pod simple-prod-query-849456d7cd-fbxv5.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:14 +0000 UTC Normal Pod simple-prod-query-849456d7cd-fbxv5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:14 +0000 UTC Normal Pod simple-prod-query-849456d7cd-fbxv5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:14 +0000 UTC Normal Pod simple-prod-query-849456d7cd-fbxv5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:14 +0000 UTC Normal Pod simple-prod-query-849456d7cd-fbxv5.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:14 +0000 UTC Normal Pod simple-prod-query-849456d7cd-fbxv5.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:14 +0000 UTC Normal Pod simple-prod-query-849456d7cd-fbxv5.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:14 +0000 UTC Normal Pod simple-prod-query-849456d7cd-jw52w.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:14 +0000 UTC Normal Pod simple-prod-query-849456d7cd-jw52w.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:14 +0000 UTC Normal Pod simple-prod-query-849456d7cd-jw52w.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:14 +0000 UTC Normal Pod simple-prod-query-849456d7cd-jw52w.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:14 +0000 UTC Normal Pod simple-prod-query-849456d7cd-jw52w.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:14 +0000 UTC Normal Pod simple-prod-query-849456d7cd-jw52w.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:14 +0000 UTC Normal Pod simple-prod-query-849456d7cd-jw52w.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:14 +0000 UTC Normal Pod simple-prod-query-849456d7cd-jw52w.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:14 +0000 UTC Normal Pod simple-prod-query-849456d7cd-jw52w.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:17 +0000 UTC Normal Pod check-span-6fltm Binding Scheduled Successfully assigned kuttl-test-present-impala/check-span-6fltm to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:17 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-6fltm job-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:17 +0000 UTC Normal Pod report-span-9n7w9 Binding Scheduled Successfully assigned kuttl-test-present-impala/report-span-9n7w9 to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:17 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-9n7w9 job-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:18 +0000 UTC Normal Pod check-span-6fltm AddedInterface Add eth0 [10.129.2.41/23] from ovn-kubernetes multus logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:18 +0000 UTC Normal Pod check-span-6fltm.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:18 +0000 UTC Normal Pod check-span-6fltm.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:18 +0000 UTC Normal Pod check-span-6fltm.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:18 +0000 UTC Normal Pod report-span-9n7w9 AddedInterface Add eth0 [10.131.0.34/23] from ovn-kubernetes multus logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:18 +0000 UTC Normal Pod report-span-9n7w9.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:18 +0000 UTC Normal Pod report-span-9n7w9.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:18 +0000 UTC Normal Pod report-span-9n7w9.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:29 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestpresentimpalasimpleprod-2-7d5c7f5744 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestpresentimpalasimpleprod-2-7d5c7mzn2t replicaset-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-2-7d5c7mzn2t Binding Scheduled Successfully assigned kuttl-test-present-impala/elasticsearch-cdm-kuttltestpresentimpalasimpleprod-2-7d5c7mzn2t to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-2-7d5c7mzn2t AddedInterface Add eth0 [10.129.2.42/23] from ovn-kubernetes multus logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-2-7d5c7mzn2t.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestpresentimpalasimpleprod-2 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestpresentimpalasimpleprod-2-7d5c7f5744 to 1 deployment-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal Pod simple-prod-collector-7884945cf7-flqzt.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal Pod simple-prod-collector-7884945cf7-scz8w.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-7884945cf7 SuccessfulDelete Deleted pod: simple-prod-collector-7884945cf7-flqzt replicaset-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-7884945cf7 SuccessfulDelete Deleted pod: simple-prod-collector-7884945cf7-scz8w replicaset-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled down replica set simple-prod-collector-7884945cf7 to 0 from 2 deployment-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal Pod simple-prod-query-849456d7cd-fbxv5.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal Pod simple-prod-query-849456d7cd-fbxv5.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal Pod simple-prod-query-849456d7cd-fbxv5.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal Pod simple-prod-query-849456d7cd-jw52w.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal Pod simple-prod-query-849456d7cd-jw52w.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal Pod simple-prod-query-849456d7cd-jw52w.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal ReplicaSet.apps simple-prod-query-849456d7cd SuccessfulDelete Deleted pod: simple-prod-query-849456d7cd-fbxv5 replicaset-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal ReplicaSet.apps simple-prod-query-849456d7cd SuccessfulDelete Deleted pod: simple-prod-query-849456d7cd-jw52w replicaset-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:31 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-849456d7cd to 0 from 2 deployment-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-2-7d5c7mzn2t.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-2-7d5c7mzn2t.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-2-7d5c7mzn2t.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-2-7d5c7mzn2t.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentimpalasimpleprod-2-7d5c7mzn2t.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:32 +0000 UTC Normal Pod simple-prod-collector-7f867c5d74-gctwf Binding Scheduled Successfully assigned kuttl-test-present-impala/simple-prod-collector-7f867c5d74-gctwf to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:32 +0000 UTC Normal Pod simple-prod-collector-7f867c5d74-kdd8g Binding Scheduled Successfully assigned kuttl-test-present-impala/simple-prod-collector-7f867c5d74-kdd8g to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:32 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-7f867c5d74 SuccessfulCreate Created pod: simple-prod-collector-7f867c5d74-kdd8g replicaset-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:32 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-7f867c5d74 SuccessfulCreate Created pod: simple-prod-collector-7f867c5d74-gctwf replicaset-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:32 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-7f867c5d74 to 2 deployment-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:32 +0000 UTC Normal Pod simple-prod-query-75d87f4559-m5qzg Binding Scheduled Successfully assigned kuttl-test-present-impala/simple-prod-query-75d87f4559-m5qzg to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:32 +0000 UTC Normal Pod simple-prod-query-75d87f4559-sspc6 Binding Scheduled Successfully assigned kuttl-test-present-impala/simple-prod-query-75d87f4559-sspc6 to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:32 +0000 UTC Normal ReplicaSet.apps simple-prod-query-75d87f4559 SuccessfulCreate Created pod: simple-prod-query-75d87f4559-sspc6 replicaset-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:32 +0000 UTC Normal ReplicaSet.apps simple-prod-query-75d87f4559 SuccessfulCreate Created pod: simple-prod-query-75d87f4559-m5qzg replicaset-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:32 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-75d87f4559 to 2 deployment-controller logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-collector-7f867c5d74-gctwf AddedInterface Add eth0 [10.129.2.43/23] from ovn-kubernetes multus logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-collector-7f867c5d74-gctwf.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-collector-7f867c5d74-gctwf.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-collector-7f867c5d74-gctwf.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-collector-7f867c5d74-kdd8g AddedInterface Add eth0 [10.131.0.35/23] from ovn-kubernetes multus logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-collector-7f867c5d74-kdd8g.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-collector-7f867c5d74-kdd8g.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-collector-7f867c5d74-kdd8g.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-query-75d87f4559-m5qzg AddedInterface Add eth0 [10.128.2.20/23] from ovn-kubernetes multus logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-query-75d87f4559-m5qzg.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-query-75d87f4559-sspc6 AddedInterface Add eth0 [10.131.0.36/23] from ovn-kubernetes multus logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-query-75d87f4559-sspc6.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-query-75d87f4559-sspc6.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-query-75d87f4559-sspc6.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-query-75d87f4559-sspc6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-query-75d87f4559-sspc6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-query-75d87f4559-sspc6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-query-75d87f4559-sspc6.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-query-75d87f4559-sspc6.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:06:35 | es-increasing-replicas | 2024-03-04 07:06:33 +0000 UTC Normal Pod simple-prod-query-75d87f4559-sspc6.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:06:35 | es-increasing-replicas | Deleting namespace: kuttl-test-present-impala === CONT kuttl/harness/es-index-cleaner-autoprov logger.go:42: 07:07:10 | es-index-cleaner-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:07:10 | es-index-cleaner-autoprov | Creating namespace: kuttl-test-cosmic-toad logger.go:42: 07:07:10 | es-index-cleaner-autoprov/1-install | starting test step 1-install logger.go:42: 07:07:10 | es-index-cleaner-autoprov/1-install | Jaeger:kuttl-test-cosmic-toad/test-es-index-cleaner-with-prefix created logger.go:42: 07:07:47 | es-index-cleaner-autoprov/1-install | test step completed 1-install logger.go:42: 07:07:47 | es-index-cleaner-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:07:47 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE test-es-index-cleaner-with-prefix /dev/null] logger.go:42: 07:07:49 | es-index-cleaner-autoprov/2-report-spans | Warning: resource jaegers/test-es-index-cleaner-with-prefix is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:07:55 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c DAYS=5 ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:07:55 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:07:56 | es-index-cleaner-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:17:56 | es-index-cleaner-autoprov/2-report-spans | test step failed 2-report-spans case.go:364: failed in step 2-report-spans case.go:366: --- Job:kuttl-test-cosmic-toad/00-report-span +++ Job:kuttl-test-cosmic-toad/00-report-span @@ -1,8 +1,162 @@ apiVersion: batch/v1 kind: Job metadata: + annotations: + kubectl.kubernetes.io/last-applied-configuration: | + {"apiVersion":"batch/v1","kind":"Job","metadata":{"annotations":{},"name":"00-report-span","namespace":"kuttl-test-cosmic-toad"},"spec":{"backoffLimit":15,"template":{"spec":{"containers":[{"command":["./reporter","--days","5","--verbose"],"env":[{"name":"JAEGER_SERVICE_NAME","value":"smoke-test-service"},{"name":"OPERATION_NAME","value":"smoke-test-operation"},{"name":"JAEGER_ENDPOINT","value":"http://test-es-index-cleaner-with-prefix-collector-headless:14268/api/traces"},{"name":"JAEGER_QUERY","value":"https://test-es-index-cleaner-with-prefix-query/api/traces"},{"name":"SECRET_PATH","value":"/var/run/secrets/api-token/token"}],"image":"registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f","name":"asserts-container","volumeMounts":[{"mountPath":"/var/run/secrets/api-token","name":"token-api-volume"}]}],"restartPolicy":"OnFailure","volumes":[{"name":"token-api-volume","secret":{"secretName":"e2e-test"}}]}}}} + labels: + batch.kubernetes.io/controller-uid: fb6778bd-8a3f-437e-a338-be9fa251a8bc + batch.kubernetes.io/job-name: 00-report-span + controller-uid: fb6778bd-8a3f-437e-a338-be9fa251a8bc + job-name: 00-report-span + managedFields: + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:kubectl.kubernetes.io/last-applied-configuration: {} + f:spec: + f:backoffLimit: {} + f:completionMode: {} + f:completions: {} + f:manualSelector: {} + f:parallelism: {} + f:podReplacementPolicy: {} + f:suspend: {} + f:template: + f:spec: + f:containers: + k:{"name":"asserts-container"}: + .: {} + f:command: {} + f:env: + .: {} + k:{"name":"JAEGER_ENDPOINT"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"JAEGER_QUERY"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"JAEGER_SERVICE_NAME"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"OPERATION_NAME"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SECRET_PATH"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:name: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/var/run/secrets/api-token"}: + .: {} + f:mountPath: {} + f:name: {} + f:dnsPolicy: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"token-api-volume"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + manager: kubectl-client-side-apply + operation: Update + time: "2024-03-04T07:07:55Z" + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:status: + f:active: {} + f:ready: {} + f:startTime: {} + f:terminating: {} + f:uncountedTerminatedPods: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-03-04T07:07:58Z" name: 00-report-span namespace: kuttl-test-cosmic-toad +spec: + backoffLimit: 15 + completionMode: NonIndexed + completions: 1 + manualSelector: false + parallelism: 1 + podReplacementPolicy: TerminatingOrFailed + selector: + matchLabels: + batch.kubernetes.io/controller-uid: fb6778bd-8a3f-437e-a338-be9fa251a8bc + suspend: false + template: + metadata: + creationTimestamp: null + labels: + batch.kubernetes.io/controller-uid: fb6778bd-8a3f-437e-a338-be9fa251a8bc + batch.kubernetes.io/job-name: 00-report-span + controller-uid: fb6778bd-8a3f-437e-a338-be9fa251a8bc + job-name: 00-report-span + spec: + containers: + - command: + - ./reporter + - --days + - "5" + - --verbose + env: + - name: JAEGER_SERVICE_NAME + value: smoke-test-service + - name: OPERATION_NAME + value: smoke-test-operation + - name: JAEGER_ENDPOINT + value: http://test-es-index-cleaner-with-prefix-collector-headless:14268/api/traces + - name: JAEGER_QUERY + value: https://test-es-index-cleaner-with-prefix-query/api/traces + - name: SECRET_PATH + value: /var/run/secrets/api-token/token + image: registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f + imagePullPolicy: IfNotPresent + name: asserts-container + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /var/run/secrets/api-token + name: token-api-volume + dnsPolicy: ClusterFirst + restartPolicy: OnFailure + schedulerName: default-scheduler + securityContext: {} + terminationGracePeriodSeconds: 30 + volumes: + - name: token-api-volume + secret: + defaultMode: 420 + secretName: e2e-test status: - succeeded: 1 + active: 1 + ready: 1 + startTime: "2024-03-04T07:07:56Z" + terminating: 0 + uncountedTerminatedPods: {} case.go:366: resource Job:kuttl-test-cosmic-toad/00-report-span: .status.succeeded: key is missing from map logger.go:42: 07:17:56 | es-index-cleaner-autoprov | es-index-cleaner-autoprov events from ns kuttl-test-cosmic-toad: logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:16 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestcosmictoadtestesindexcleane-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestcosmictoadtestesindexcleane-1-fb948967d to 1 deployment-controller logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcosmictoadtestesindexcleane-1-f4vnwm Binding Scheduled Successfully assigned kuttl-test-cosmic-toad/elasticsearch-cdm-kuttltestcosmictoadtestesindexcleane-1-f4vnwm to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcosmictoadtestesindexcleane-1-f4vnwm AddedInterface Add eth0 [10.128.2.21/23] from ovn-kubernetes multus logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcosmictoadtestesindexcleane-1-f4vnwm.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcosmictoadtestesindexcleane-1-f4vnwm.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcosmictoadtestesindexcleane-1-f4vnwm.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcosmictoadtestesindexcleane-1-f4vnwm.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcosmictoadtestesindexcleane-1-f4vnwm.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcosmictoadtestesindexcleane-1-f4vnwm.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:17 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestcosmictoadtestesindexcleane-1-fb948967d SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestcosmictoadtestesindexcleane-1-f4vnwm replicaset-controller logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:27 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcosmictoadtestesindexcleane-1-f4vnwm.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:32 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcosmictoadtestesindexcleane-1-f4vnwm.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-757c5d7d8d-nf9xh Binding Scheduled Successfully assigned kuttl-test-cosmic-toad/test-es-index-cleaner-with-prefix-collector-757c5d7d8d-nf9xh to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:43 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-collector-757c5d7d8d SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-collector-757c5d7d8d-nf9xh replicaset-controller logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:43 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-collector ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-collector-757c5d7d8d to 1 deployment-controller logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f4cfb67d-fqb9w Binding Scheduled Successfully assigned kuttl-test-cosmic-toad/test-es-index-cleaner-with-prefix-query-64f4cfb67d-fqb9w to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:43 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-64f4cfb67d SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-64f4cfb67d-fqb9w replicaset-controller logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:43 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-64f4cfb67d to 1 deployment-controller logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:44 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-757c5d7d8d-nf9xh AddedInterface Add eth0 [10.129.2.44/23] from ovn-kubernetes multus logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:44 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-757c5d7d8d-nf9xh.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" already present on machine kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:44 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-757c5d7d8d-nf9xh.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:44 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-757c5d7d8d-nf9xh.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:44 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f4cfb67d-fqb9w AddedInterface Add eth0 [10.131.0.37/23] from ovn-kubernetes multus logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:44 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f4cfb67d-fqb9w.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:44 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f4cfb67d-fqb9w.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:44 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f4cfb67d-fqb9w.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:44 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f4cfb67d-fqb9w.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:44 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f4cfb67d-fqb9w.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:44 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f4cfb67d-fqb9w.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:44 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f4cfb67d-fqb9w.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:44 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f4cfb67d-fqb9w.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:44 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f4cfb67d-fqb9w.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:56 +0000 UTC Normal Pod 00-report-span-z28v5 Binding Scheduled Successfully assigned kuttl-test-cosmic-toad/00-report-span-z28v5 to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:56 +0000 UTC Normal Pod 00-report-span-z28v5 AddedInterface Add eth0 [10.129.2.45/23] from ovn-kubernetes multus logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:56 +0000 UTC Normal Pod 00-report-span-z28v5.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:56 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-z28v5 job-controller logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:57 +0000 UTC Normal Pod 00-report-span-z28v5.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:57 +0000 UTC Normal Pod 00-report-span-z28v5.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:59 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:59 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:07:59 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:08:44 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:08:44 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-757c5d7d8d-nf9xh horizontal-pod-autoscaler logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:08:44 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:17:56 | es-index-cleaner-autoprov | 2024-03-04 07:12:59 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-757c5d7d8d-nf9xh horizontal-pod-autoscaler logger.go:42: 07:17:56 | es-index-cleaner-autoprov | Deleting namespace: kuttl-test-cosmic-toad === CONT kuttl/harness/es-from-aio-to-production logger.go:42: 07:18:03 | es-from-aio-to-production | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:18:03 | es-from-aio-to-production | Creating namespace: kuttl-test-arriving-treefrog logger.go:42: 07:18:03 | es-from-aio-to-production/0-install | starting test step 0-install logger.go:42: 07:18:03 | es-from-aio-to-production/0-install | Jaeger:kuttl-test-arriving-treefrog/my-jaeger created logger.go:42: 07:18:11 | es-from-aio-to-production/0-install | test step completed 0-install logger.go:42: 07:18:11 | es-from-aio-to-production/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:18:11 | es-from-aio-to-production/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:18:12 | es-from-aio-to-production/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:18:19 | es-from-aio-to-production/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:18:19 | es-from-aio-to-production/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:18:19 | es-from-aio-to-production/1-smoke-test | job.batch/report-span created logger.go:42: 07:18:19 | es-from-aio-to-production/1-smoke-test | job.batch/check-span created logger.go:42: 07:18:32 | es-from-aio-to-production/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:18:32 | es-from-aio-to-production/3-install | starting test step 3-install logger.go:42: 07:18:32 | es-from-aio-to-production/3-install | Jaeger:kuttl-test-arriving-treefrog/my-jaeger updated logger.go:42: 07:19:06 | es-from-aio-to-production/3-install | test step completed 3-install logger.go:42: 07:19:06 | es-from-aio-to-production/4-smoke-test | starting test step 4-smoke-test logger.go:42: 07:19:06 | es-from-aio-to-production/4-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:19:13 | es-from-aio-to-production/4-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:19:14 | es-from-aio-to-production/4-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:19:14 | es-from-aio-to-production/4-smoke-test | job.batch/report-span unchanged logger.go:42: 07:19:14 | es-from-aio-to-production/4-smoke-test | job.batch/check-span unchanged logger.go:42: 07:19:14 | es-from-aio-to-production/4-smoke-test | test step completed 4-smoke-test logger.go:42: 07:19:14 | es-from-aio-to-production | es-from-aio-to-production events from ns kuttl-test-arriving-treefrog: logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:07 +0000 UTC Normal Pod my-jaeger-6ffdb89d56-mc56w Binding Scheduled Successfully assigned kuttl-test-arriving-treefrog/my-jaeger-6ffdb89d56-mc56w to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:07 +0000 UTC Normal Pod my-jaeger-6ffdb89d56-mc56w AddedInterface Add eth0 [10.128.2.22/23] from ovn-kubernetes multus logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:07 +0000 UTC Normal Pod my-jaeger-6ffdb89d56-mc56w.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:07 +0000 UTC Normal ReplicaSet.apps my-jaeger-6ffdb89d56 SuccessfulCreate Created pod: my-jaeger-6ffdb89d56-mc56w replicaset-controller logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:07 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6ffdb89d56 to 1 deployment-controller logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:09 +0000 UTC Normal Pod my-jaeger-6ffdb89d56-mc56w.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" in 1.883s (1.883s including waiting) kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:09 +0000 UTC Normal Pod my-jaeger-6ffdb89d56-mc56w.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:09 +0000 UTC Normal Pod my-jaeger-6ffdb89d56-mc56w.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:09 +0000 UTC Normal Pod my-jaeger-6ffdb89d56-mc56w.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:09 +0000 UTC Normal Pod my-jaeger-6ffdb89d56-mc56w.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:09 +0000 UTC Normal Pod my-jaeger-6ffdb89d56-mc56w.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:14 +0000 UTC Normal Pod my-jaeger-6ffdb89d56-mc56w.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:14 +0000 UTC Normal Pod my-jaeger-6ffdb89d56-mc56w.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:14 +0000 UTC Normal ReplicaSet.apps my-jaeger-6ffdb89d56 SuccessfulDelete Deleted pod: my-jaeger-6ffdb89d56-mc56w replicaset-controller logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:14 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-6ffdb89d56 to 0 from 1 deployment-controller logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:15 +0000 UTC Normal Pod my-jaeger-c86cd79d8-gcncl Binding Scheduled Successfully assigned kuttl-test-arriving-treefrog/my-jaeger-c86cd79d8-gcncl to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:15 +0000 UTC Normal ReplicaSet.apps my-jaeger-c86cd79d8 SuccessfulCreate Created pod: my-jaeger-c86cd79d8-gcncl replicaset-controller logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:15 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-c86cd79d8 to 1 deployment-controller logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:16 +0000 UTC Normal Pod my-jaeger-c86cd79d8-gcncl AddedInterface Add eth0 [10.128.2.23/23] from ovn-kubernetes multus logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:16 +0000 UTC Normal Pod my-jaeger-c86cd79d8-gcncl.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:16 +0000 UTC Normal Pod my-jaeger-c86cd79d8-gcncl.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:16 +0000 UTC Normal Pod my-jaeger-c86cd79d8-gcncl.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:16 +0000 UTC Normal Pod my-jaeger-c86cd79d8-gcncl.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:16 +0000 UTC Normal Pod my-jaeger-c86cd79d8-gcncl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:16 +0000 UTC Normal Pod my-jaeger-c86cd79d8-gcncl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:19 +0000 UTC Normal Pod check-span-zqzm5 Binding Scheduled Successfully assigned kuttl-test-arriving-treefrog/check-span-zqzm5 to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:19 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-zqzm5 job-controller logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:19 +0000 UTC Normal Pod report-span-frpcf Binding Scheduled Successfully assigned kuttl-test-arriving-treefrog/report-span-frpcf to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:19 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-frpcf job-controller logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:20 +0000 UTC Normal Pod check-span-zqzm5 AddedInterface Add eth0 [10.131.0.38/23] from ovn-kubernetes multus logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:20 +0000 UTC Normal Pod check-span-zqzm5.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:20 +0000 UTC Normal Pod check-span-zqzm5.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:20 +0000 UTC Normal Pod check-span-zqzm5.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:20 +0000 UTC Normal Pod report-span-frpcf AddedInterface Add eth0 [10.129.2.47/23] from ovn-kubernetes multus logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:20 +0000 UTC Normal Pod report-span-frpcf.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:20 +0000 UTC Normal Pod report-span-frpcf.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:20 +0000 UTC Normal Pod report-span-frpcf.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:31 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestarrivingtreefrogmyjaeger-1-67c67gg6d Binding Scheduled Successfully assigned kuttl-test-arriving-treefrog/elasticsearch-cdm-kuttltestarrivingtreefrogmyjaeger-1-67c67gg6d to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestarrivingtreefrogmyjaeger-1-67c67gg6d AddedInterface Add eth0 [10.128.2.24/23] from ovn-kubernetes multus logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestarrivingtreefrogmyjaeger-1-67c67gg6d.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestarrivingtreefrogmyjaeger-1-67c67gg6d.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestarrivingtreefrogmyjaeger-1-67c67gg6d.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestarrivingtreefrogmyjaeger-1-67c67gg6d.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestarrivingtreefrogmyjaeger-1-67c67gg6d.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestarrivingtreefrogmyjaeger-1-67c67gg6d.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:36 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestarrivingtreefrogmyjaeger-1-67c6cc79dd SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestarrivingtreefrogmyjaeger-1-67c67gg6d replicaset-controller logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:36 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestarrivingtreefrogmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestarrivingtreefrogmyjaeger-1-67c6cc79dd to 1 deployment-controller logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:46 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestarrivingtreefrogmyjaeger-1-67c67gg6d.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:51 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestarrivingtreefrogmyjaeger-1-67c67gg6d.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:18:53 +0000 UTC Normal Job.batch report-span Completed Job completed job-controller logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:02 +0000 UTC Normal Pod my-jaeger-collector-59b4d446fb-g82b6 Binding Scheduled Successfully assigned kuttl-test-arriving-treefrog/my-jaeger-collector-59b4d446fb-g82b6 to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:02 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-59b4d446fb SuccessfulCreate Created pod: my-jaeger-collector-59b4d446fb-g82b6 replicaset-controller logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:02 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-59b4d446fb to 1 deployment-controller logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:02 +0000 UTC Normal Pod my-jaeger-query-5dccb54787-69tkx Binding Scheduled Successfully assigned kuttl-test-arriving-treefrog/my-jaeger-query-5dccb54787-69tkx to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:02 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5dccb54787 SuccessfulCreate Created pod: my-jaeger-query-5dccb54787-69tkx replicaset-controller logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:02 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-5dccb54787 to 1 deployment-controller logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:03 +0000 UTC Normal Pod my-jaeger-c86cd79d8-gcncl.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:03 +0000 UTC Normal Pod my-jaeger-c86cd79d8-gcncl.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:03 +0000 UTC Normal Pod my-jaeger-collector-59b4d446fb-g82b6 AddedInterface Add eth0 [10.129.2.48/23] from ovn-kubernetes multus logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:03 +0000 UTC Normal Pod my-jaeger-collector-59b4d446fb-g82b6.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" already present on machine kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:03 +0000 UTC Normal Pod my-jaeger-collector-59b4d446fb-g82b6.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:03 +0000 UTC Normal Pod my-jaeger-collector-59b4d446fb-g82b6.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:03 +0000 UTC Normal Pod my-jaeger-query-5dccb54787-69tkx AddedInterface Add eth0 [10.131.0.39/23] from ovn-kubernetes multus logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:03 +0000 UTC Normal Pod my-jaeger-query-5dccb54787-69tkx.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:03 +0000 UTC Normal Pod my-jaeger-query-5dccb54787-69tkx.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:03 +0000 UTC Normal Pod my-jaeger-query-5dccb54787-69tkx.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:03 +0000 UTC Normal Pod my-jaeger-query-5dccb54787-69tkx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:03 +0000 UTC Normal Pod my-jaeger-query-5dccb54787-69tkx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:03 +0000 UTC Normal Pod my-jaeger-query-5dccb54787-69tkx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:03 +0000 UTC Normal Pod my-jaeger-query-5dccb54787-69tkx.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:03 +0000 UTC Normal Pod my-jaeger-query-5dccb54787-69tkx.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | 2024-03-04 07:19:03 +0000 UTC Normal Pod my-jaeger-query-5dccb54787-69tkx.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:19:14 | es-from-aio-to-production | Deleting namespace: kuttl-test-arriving-treefrog === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (1186.73s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.64s) --- PASS: kuttl/harness/es-multiinstance (118.43s) --- PASS: kuttl/harness/es-simple-prod (5.99s) --- PASS: kuttl/harness/es-rollover-autoprov (221.21s) --- PASS: kuttl/harness/es-increasing-replicas (103.55s) --- FAIL: kuttl/harness/es-index-cleaner-autoprov (653.14s) --- PASS: kuttl/harness/es-from-aio-to-production (77.72s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name elasticsearch --report --output /logs/artifacts/elasticsearch.xml ./artifacts/kuttl-report.xml time="2024-03-04T07:19:22Z" level=debug msg="Setting a new name for the test suites" time="2024-03-04T07:19:22Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-03-04T07:19:22Z" level=debug msg="normalizing test case names" time="2024-03-04T07:19:22Z" level=debug msg="elasticsearch/artifacts -> elasticsearch_artifacts" time="2024-03-04T07:19:22Z" level=debug msg="elasticsearch/es-multiinstance -> elasticsearch_es_multiinstance" time="2024-03-04T07:19:22Z" level=debug msg="elasticsearch/es-simple-prod -> elasticsearch_es_simple_prod" time="2024-03-04T07:19:22Z" level=debug msg="elasticsearch/es-rollover-autoprov -> elasticsearch_es_rollover_autoprov" time="2024-03-04T07:19:22Z" level=debug msg="elasticsearch/es-increasing-replicas -> elasticsearch_es_increasing_replicas" time="2024-03-04T07:19:22Z" level=debug msg="elasticsearch/es-index-cleaner-autoprov -> elasticsearch_es_index_cleaner_autoprov" time="2024-03-04T07:19:22Z" level=debug msg="elasticsearch/es-from-aio-to-production -> elasticsearch_es_from_aio_to_production" +-----------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------+--------+ | elasticsearch_artifacts | passed | | elasticsearch_es_multiinstance | passed | | elasticsearch_es_simple_prod | passed | | elasticsearch_es_rollover_autoprov | passed | | elasticsearch_es_increasing_replicas | passed | | elasticsearch_es_index_cleaner_autoprov | failed | | elasticsearch_es_from_aio_to_production | passed | +-----------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh examples false true + '[' 3 -ne 3 ']' + test_suite_name=examples + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/examples.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-examples make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ VERTX_IMG=jaegertracing/vertx-create-span:operator-e2e-tests \ ./tests/e2e/examples/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-02-29-062601 True False 31m Cluster version is 4.16.0-0.nightly-2024-02-29-062601' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-02-29-062601 True False 31m Cluster version is 4.16.0-0.nightly-2024-02-29-062601' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/examples/render.sh ++ export SUITE_DIR=./tests/e2e/examples ++ SUITE_DIR=./tests/e2e/examples ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/examples ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test examples-agent-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-agent-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-with-priority-class\e[0m' Rendering files for test examples-agent-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + mkdir -p examples-agent-with-priority-class + cd examples-agent-with-priority-class + example_name=agent-with-priority-class + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + '[' true '!=' true ']' + render_install_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + install_file=./02-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/agent-with-priority-class.yaml -o ./02-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./02-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./02-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./02-install.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + JAEGER_NAME=agent-as-daemonset + local jaeger_strategy ++ get_jaeger_strategy ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./02-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./02-install.yaml ++ strategy=DaemonSet ++ '[' DaemonSet = null ']' ++ echo DaemonSet ++ return 0 + jaeger_strategy=DaemonSet + '[' DaemonSet = DaemonSet ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./02-assert.yaml + render_smoke_test_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + jaeger_name=agent-as-daemonset + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test agent-as-daemonset true 02 + '[' 3 -ne 3 ']' + jaeger=agent-as-daemonset + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-all-in-one-with-options + '[' 1 -ne 1 ']' + test_name=examples-all-in-one-with-options + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-all-in-one-with-options' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-all-in-one-with-options\e[0m' Rendering files for test examples-all-in-one-with-options + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-with-priority-class + '[' examples-agent-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-all-in-one-with-options + cd examples-all-in-one-with-options + example_name=all-in-one-with-options + render_install_example all-in-one-with-options 00 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/all-in-one-with-options.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + JAEGER_NAME=my-jaeger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.metadata.name="my-jaeger"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i 'del(.spec.allInOne.image)' ./00-install.yaml + render_smoke_test_example all-in-one-with-options 01 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + jaeger_name=my-jaeger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + sed -i s~my-jaeger-query:443~my-jaeger-query:443/jaeger~gi ./01-smoke-test.yaml + start_test examples-business-application-injected-sidecar + '[' 1 -ne 1 ']' + test_name=examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-business-application-injected-sidecar' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-business-application-injected-sidecar\e[0m' Rendering files for test examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-all-in-one-with-options + '[' examples-all-in-one-with-options '!=' _build ']' + cd .. + mkdir -p examples-business-application-injected-sidecar + cd examples-business-application-injected-sidecar + example_name=simplest + cp /tmp/jaeger-tests/examples/business-application-injected-sidecar.yaml ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].image=strenv(VERTX_IMG)' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.path="/"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.port=8080' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.initialDelaySeconds=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.failureThreshold=3' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.periodSeconds=10' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.successThreshold=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.timeoutSeconds=1' ./00-install.yaml + render_install_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example simplest 02 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 02 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-collector-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-collector-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-collector-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-collector-with-priority-class\e[0m' Rendering files for test examples-collector-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-business-application-injected-sidecar + '[' examples-business-application-injected-sidecar '!=' _build ']' + cd .. + mkdir -p examples-collector-with-priority-class + cd examples-collector-with-priority-class + example_name=collector-with-priority-class + render_install_example collector-with-priority-class 00 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/collector-with-priority-class.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + JAEGER_NAME=collector-with-high-priority + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example collector-with-priority-class 01 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + jaeger_name=collector-with-high-priority + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test collector-with-high-priority true 01 + '[' 3 -ne 3 ']' + jaeger=collector-with-high-priority + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + export JAEGER_NAME=collector-with-high-priority + JAEGER_NAME=collector-with-high-priority + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-service-types + '[' 1 -ne 1 ']' + test_name=examples-service-types + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-service-types' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-service-types\e[0m' Rendering files for test examples-service-types + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-collector-with-priority-class + '[' examples-collector-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-service-types + cd examples-service-types + example_name=service-types + render_install_example service-types 00 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/service-types.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + JAEGER_NAME=service-types + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example service-types 01 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/service-types.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/service-types.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/service-types.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/service-types.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + jaeger_name=service-types + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test service-types true 01 + '[' 3 -ne 3 ']' + jaeger=service-types + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + export JAEGER_NAME=service-types + JAEGER_NAME=service-types + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod + '[' 1 -ne 1 ']' + test_name=examples-simple-prod + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod\e[0m' Rendering files for test examples-simple-prod + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-service-types + '[' examples-service-types '!=' _build ']' + cd .. + mkdir -p examples-simple-prod + cd examples-simple-prod + example_name=simple-prod + render_install_example simple-prod 01 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod 02 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod-with-volumes + '[' 1 -ne 1 ']' + test_name=examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod-with-volumes' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod-with-volumes\e[0m' Rendering files for test examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod + '[' examples-simple-prod '!=' _build ']' + cd .. + mkdir -p examples-simple-prod-with-volumes + cd examples-simple-prod-with-volumes + example_name=simple-prod-with-volumes + render_install_example simple-prod-with-volumes 01 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod-with-volumes 02 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + /tmp/jaeger-tests/bin/gomplate -f ./03-check-volume.yaml.template -o 03-check-volume.yaml + start_test examples-simplest + '[' 1 -ne 1 ']' + test_name=examples-simplest + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simplest' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simplest\e[0m' Rendering files for test examples-simplest + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod-with-volumes + '[' examples-simple-prod-with-volumes '!=' _build ']' + cd .. + mkdir -p examples-simplest + cd examples-simplest + example_name=simplest + render_install_example simplest 00 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 01 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger + '[' 1 -ne 1 ']' + test_name=examples-with-badger + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger\e[0m' Rendering files for test examples-with-badger + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simplest + '[' examples-simplest '!=' _build ']' + cd .. + mkdir -p examples-with-badger + cd examples-with-badger + example_name=with-badger + render_install_example with-badger 00 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + JAEGER_NAME=with-badger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger 01 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + jaeger_name=with-badger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + export JAEGER_NAME=with-badger + JAEGER_NAME=with-badger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger-and-volume + '[' 1 -ne 1 ']' + test_name=examples-with-badger-and-volume + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger-and-volume' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger-and-volume\e[0m' Rendering files for test examples-with-badger-and-volume + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger + '[' examples-with-badger '!=' _build ']' + cd .. + mkdir -p examples-with-badger-and-volume + cd examples-with-badger-and-volume + example_name=with-badger-and-volume + render_install_example with-badger-and-volume 00 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger-and-volume.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + JAEGER_NAME=with-badger-and-volume + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger-and-volume 01 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + jaeger_name=with-badger-and-volume + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger-and-volume true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger-and-volume + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + export JAEGER_NAME=with-badger-and-volume + JAEGER_NAME=with-badger-and-volume + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-cassandra + '[' 1 -ne 1 ']' + test_name=examples-with-cassandra + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-cassandra' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-cassandra\e[0m' Rendering files for test examples-with-cassandra + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger-and-volume + '[' examples-with-badger-and-volume '!=' _build ']' + cd .. + mkdir -p examples-with-cassandra + cd examples-with-cassandra + example_name=with-cassandra + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-cassandra 01 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-cassandra.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + JAEGER_NAME=with-cassandra + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-cassandra 02 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-cassandra.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-cassandra.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + jaeger_name=with-cassandra + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-cassandra true 02 + '[' 3 -ne 3 ']' + jaeger=with-cassandra + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + export JAEGER_NAME=with-cassandra + JAEGER_NAME=with-cassandra + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-sampling + '[' 1 -ne 1 ']' + test_name=examples-with-sampling + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-sampling' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-sampling\e[0m' Rendering files for test examples-with-sampling + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-cassandra + '[' examples-with-cassandra '!=' _build ']' + cd .. + mkdir -p examples-with-sampling + cd examples-with-sampling + export example_name=with-sampling + example_name=with-sampling + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-sampling 01 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-sampling.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + JAEGER_NAME=with-sampling + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-sampling 02 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-sampling.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-sampling.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + jaeger_name=with-sampling + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-sampling true 02 + '[' 3 -ne 3 ']' + jaeger=with-sampling + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + export JAEGER_NAME=with-sampling + JAEGER_NAME=with-sampling + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-agent-as-daemonset + '[' 1 -ne 1 ']' + test_name=examples-agent-as-daemonset + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-as-daemonset' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-as-daemonset\e[0m' Rendering files for test examples-agent-as-daemonset + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-sampling + '[' examples-with-sampling '!=' _build ']' + cd .. + mkdir -p examples-agent-as-daemonset + cd examples-agent-as-daemonset + '[' true = true ']' + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/agent-as-daemonset.yaml -o 02-install.yaml + '[' true = true ']' + start_test examples-openshift-with-htpasswd + '[' 1 -ne 1 ']' + test_name=examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-openshift-with-htpasswd' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-openshift-with-htpasswd\e[0m' Rendering files for test examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-as-daemonset + '[' examples-agent-as-daemonset '!=' _build ']' + cd .. + mkdir -p examples-openshift-with-htpasswd + cd examples-openshift-with-htpasswd + export JAEGER_NAME=with-htpasswd + JAEGER_NAME=with-htpasswd + export JAEGER_USERNAME=awesomeuser + JAEGER_USERNAME=awesomeuser + export JAEGER_PASSWORD=awesomepassword + JAEGER_PASSWORD=awesomepassword + export 'JAEGER_USER_PASSWORD_HASH=awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' + JAEGER_USER_PASSWORD_HASH='awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ echo 'awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ base64 + SECRET=YXdlc29tZXVzZXI6e1NIQX11VWRxUFZVeXFOQm1FUlUwUXhqM0tGYVpuanc9Cg== + /tmp/jaeger-tests/bin/gomplate -f ./00-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/with-htpasswd.yaml -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + export 'GET_URL_COMMAND=kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + export 'URL=https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + INSECURE=true + JAEGER_USERNAME= + JAEGER_PASSWORD= + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-unsecured.yaml + JAEGER_USERNAME=wronguser + JAEGER_PASSWORD=wrongpassword + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-check-unauthorized.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./04-check-authorized.yaml + skip_test examples-agent-as-daemonset 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-agent-as-daemonset + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-openshift-with-htpasswd + '[' examples-openshift-with-htpasswd '!=' _build ']' + cd .. + rm -rf examples-agent-as-daemonset + warning 'examples-agent-as-daemonset: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-agent-as-daemonset: This test is flaky in Prow CI\e[0m' WAR: examples-agent-as-daemonset: This test is flaky in Prow CI + skip_test examples-with-badger-and-volume 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-with-badger-and-volume + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-with-badger-and-volume + warning 'examples-with-badger-and-volume: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-with-badger-and-volume: This test is flaky in Prow CI\e[0m' WAR: examples-with-badger-and-volume: This test is flaky in Prow CI + skip_test examples-collector-with-priority-class 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-collector-with-priority-class + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-collector-with-priority-class + warning 'examples-collector-with-priority-class: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-collector-with-priority-class: This test is flaky in Prow CI\e[0m' WAR: examples-collector-with-priority-class: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running examples E2E tests' Running examples E2E tests + cd tests/e2e/examples/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2507913876 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 12 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/examples-agent-with-priority-class === PAUSE kuttl/harness/examples-agent-with-priority-class === RUN kuttl/harness/examples-all-in-one-with-options === PAUSE kuttl/harness/examples-all-in-one-with-options === RUN kuttl/harness/examples-business-application-injected-sidecar === PAUSE kuttl/harness/examples-business-application-injected-sidecar === RUN kuttl/harness/examples-openshift-with-htpasswd === PAUSE kuttl/harness/examples-openshift-with-htpasswd === RUN kuttl/harness/examples-service-types === PAUSE kuttl/harness/examples-service-types === RUN kuttl/harness/examples-simple-prod === PAUSE kuttl/harness/examples-simple-prod === RUN kuttl/harness/examples-simple-prod-with-volumes === PAUSE kuttl/harness/examples-simple-prod-with-volumes === RUN kuttl/harness/examples-simplest === PAUSE kuttl/harness/examples-simplest === RUN kuttl/harness/examples-with-badger === PAUSE kuttl/harness/examples-with-badger === RUN kuttl/harness/examples-with-cassandra === PAUSE kuttl/harness/examples-with-cassandra === RUN kuttl/harness/examples-with-sampling === PAUSE kuttl/harness/examples-with-sampling === CONT kuttl/harness/artifacts logger.go:42: 07:19:53 | artifacts | Creating namespace: kuttl-test-engaged-hermit logger.go:42: 07:19:53 | artifacts | artifacts events from ns kuttl-test-engaged-hermit: logger.go:42: 07:19:53 | artifacts | Deleting namespace: kuttl-test-engaged-hermit === CONT kuttl/harness/examples-simple-prod logger.go:42: 07:19:59 | examples-simple-prod | Creating namespace: kuttl-test-becoming-killdeer logger.go:42: 07:19:59 | examples-simple-prod/1-install | starting test step 1-install logger.go:42: 07:19:59 | examples-simple-prod/1-install | Jaeger:kuttl-test-becoming-killdeer/simple-prod created logger.go:42: 07:20:35 | examples-simple-prod/1-install | test step completed 1-install logger.go:42: 07:20:35 | examples-simple-prod/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:20:35 | examples-simple-prod/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:20:37 | examples-simple-prod/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:20:43 | examples-simple-prod/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:20:44 | examples-simple-prod/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:20:44 | examples-simple-prod/2-smoke-test | job.batch/report-span created logger.go:42: 07:20:44 | examples-simple-prod/2-smoke-test | job.batch/check-span created logger.go:42: 07:20:55 | examples-simple-prod/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:20:55 | examples-simple-prod | examples-simple-prod events from ns kuttl-test-becoming-killdeer: logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:06 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestbecomingkilldeersimpleprod-1-7d4d8cf788 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestbecomingkilldeersimpleprod-1-7dnfnsp replicaset-controller logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingkilldeersimpleprod-1-7dnfnsp Binding Scheduled Successfully assigned kuttl-test-becoming-killdeer/elasticsearch-cdm-kuttltestbecomingkilldeersimpleprod-1-7dnfnsp to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingkilldeersimpleprod-1-7dnfnsp AddedInterface Add eth0 [10.128.2.25/23] from ovn-kubernetes multus logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingkilldeersimpleprod-1-7dnfnsp.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingkilldeersimpleprod-1-7dnfnsp.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingkilldeersimpleprod-1-7dnfnsp.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingkilldeersimpleprod-1-7dnfnsp.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingkilldeersimpleprod-1-7dnfnsp.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingkilldeersimpleprod-1-7dnfnsp.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:06 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestbecomingkilldeersimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestbecomingkilldeersimpleprod-1-7d4d8cf788 to 1 deployment-controller logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:16 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestbecomingkilldeersimpleprod-1-7dnfnsp.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:21 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestbecomingkilldeersimpleprod-1-7dnfnsp.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:32 +0000 UTC Normal Pod simple-prod-collector-568c6974cb-268nm Binding Scheduled Successfully assigned kuttl-test-becoming-killdeer/simple-prod-collector-568c6974cb-268nm to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:32 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-568c6974cb SuccessfulCreate Created pod: simple-prod-collector-568c6974cb-268nm replicaset-controller logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:32 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-568c6974cb to 1 deployment-controller logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:32 +0000 UTC Normal ReplicaSet.apps simple-prod-query-766d68b5b5 SuccessfulCreate Created pod: simple-prod-query-766d68b5b5-4jzht replicaset-controller logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:32 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-766d68b5b5 to 1 deployment-controller logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:33 +0000 UTC Normal Pod simple-prod-collector-568c6974cb-268nm AddedInterface Add eth0 [10.129.2.49/23] from ovn-kubernetes multus logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:33 +0000 UTC Normal Pod simple-prod-collector-568c6974cb-268nm.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" already present on machine kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:33 +0000 UTC Normal Pod simple-prod-query-766d68b5b5-4jzht Binding Scheduled Successfully assigned kuttl-test-becoming-killdeer/simple-prod-query-766d68b5b5-4jzht to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:33 +0000 UTC Normal Pod simple-prod-query-766d68b5b5-4jzht AddedInterface Add eth0 [10.131.0.40/23] from ovn-kubernetes multus logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:33 +0000 UTC Normal Pod simple-prod-query-766d68b5b5-4jzht.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:33 +0000 UTC Normal Pod simple-prod-query-766d68b5b5-4jzht.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:33 +0000 UTC Normal Pod simple-prod-query-766d68b5b5-4jzht.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:33 +0000 UTC Normal Pod simple-prod-query-766d68b5b5-4jzht.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:33 +0000 UTC Normal Pod simple-prod-query-766d68b5b5-4jzht.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:33 +0000 UTC Normal Pod simple-prod-query-766d68b5b5-4jzht.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:33 +0000 UTC Normal Pod simple-prod-query-766d68b5b5-4jzht.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:33 +0000 UTC Normal Pod simple-prod-query-766d68b5b5-4jzht.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:34 +0000 UTC Normal Pod simple-prod-collector-568c6974cb-268nm.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:34 +0000 UTC Normal Pod simple-prod-collector-568c6974cb-268nm.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:34 +0000 UTC Normal Pod simple-prod-query-766d68b5b5-4jzht.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:39 +0000 UTC Normal Pod simple-prod-query-766d68b5b5-4jzht.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:39 +0000 UTC Normal Pod simple-prod-query-766d68b5b5-4jzht.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:39 +0000 UTC Normal Pod simple-prod-query-766d68b5b5-4jzht.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:39 +0000 UTC Normal ReplicaSet.apps simple-prod-query-766d68b5b5 SuccessfulDelete Deleted pod: simple-prod-query-766d68b5b5-4jzht replicaset-controller logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:39 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-766d68b5b5 to 0 from 1 deployment-controller logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:40 +0000 UTC Normal Pod simple-prod-query-6886c98b6b-d4xlm Binding Scheduled Successfully assigned kuttl-test-becoming-killdeer/simple-prod-query-6886c98b6b-d4xlm to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:40 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6886c98b6b SuccessfulCreate Created pod: simple-prod-query-6886c98b6b-d4xlm replicaset-controller logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:40 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6886c98b6b to 1 deployment-controller logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:41 +0000 UTC Normal Pod simple-prod-query-6886c98b6b-d4xlm AddedInterface Add eth0 [10.131.0.41/23] from ovn-kubernetes multus logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:41 +0000 UTC Normal Pod simple-prod-query-6886c98b6b-d4xlm.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:41 +0000 UTC Normal Pod simple-prod-query-6886c98b6b-d4xlm.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:41 +0000 UTC Normal Pod simple-prod-query-6886c98b6b-d4xlm.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:41 +0000 UTC Normal Pod simple-prod-query-6886c98b6b-d4xlm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:41 +0000 UTC Normal Pod simple-prod-query-6886c98b6b-d4xlm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:41 +0000 UTC Normal Pod simple-prod-query-6886c98b6b-d4xlm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:41 +0000 UTC Normal Pod simple-prod-query-6886c98b6b-d4xlm.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:41 +0000 UTC Normal Pod simple-prod-query-6886c98b6b-d4xlm.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:41 +0000 UTC Normal Pod simple-prod-query-6886c98b6b-d4xlm.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:44 +0000 UTC Normal Pod check-span-4q2jk Binding Scheduled Successfully assigned kuttl-test-becoming-killdeer/check-span-4q2jk to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:44 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-4q2jk job-controller logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:44 +0000 UTC Normal Pod report-span-4szrv Binding Scheduled Successfully assigned kuttl-test-becoming-killdeer/report-span-4szrv to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:44 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-4szrv job-controller logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:45 +0000 UTC Normal Pod check-span-4q2jk AddedInterface Add eth0 [10.129.2.51/23] from ovn-kubernetes multus logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:45 +0000 UTC Normal Pod check-span-4q2jk.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:45 +0000 UTC Normal Pod check-span-4q2jk.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:45 +0000 UTC Normal Pod check-span-4q2jk.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:45 +0000 UTC Normal Pod report-span-4szrv AddedInterface Add eth0 [10.129.2.50/23] from ovn-kubernetes multus logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:45 +0000 UTC Normal Pod report-span-4szrv.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:45 +0000 UTC Normal Pod report-span-4szrv.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:45 +0000 UTC Normal Pod report-span-4szrv.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:51 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:51 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:51 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:20:55 | examples-simple-prod | 2024-03-04 07:20:55 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:20:55 | examples-simple-prod | Deleting namespace: kuttl-test-becoming-killdeer === CONT kuttl/harness/examples-with-sampling logger.go:42: 07:21:07 | examples-with-sampling | Creating namespace: kuttl-test-renewed-hare logger.go:42: 07:21:07 | examples-with-sampling/0-install | starting test step 0-install logger.go:42: 07:21:07 | examples-with-sampling/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:21:08 | examples-with-sampling/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:21:08 | examples-with-sampling/0-install | >>>> Creating namespace kuttl-test-renewed-hare logger.go:42: 07:21:08 | examples-with-sampling/0-install | kubectl create namespace kuttl-test-renewed-hare 2>&1 | grep -v "already exists" || true logger.go:42: 07:21:08 | examples-with-sampling/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-renewed-hare 2>&1 | grep -v "already exists" || true logger.go:42: 07:21:08 | examples-with-sampling/0-install | service/cassandra created logger.go:42: 07:21:08 | examples-with-sampling/0-install | statefulset.apps/cassandra created logger.go:42: 07:21:08 | examples-with-sampling/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:21:18 | examples-with-sampling/0-install | test step completed 0-install logger.go:42: 07:21:18 | examples-with-sampling/1-install | starting test step 1-install logger.go:42: 07:21:18 | examples-with-sampling/1-install | Jaeger:kuttl-test-renewed-hare/with-sampling created logger.go:42: 07:21:23 | examples-with-sampling/1-install | test step completed 1-install logger.go:42: 07:21:23 | examples-with-sampling/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:21:23 | examples-with-sampling/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-sampling /dev/null] logger.go:42: 07:21:24 | examples-with-sampling/2-smoke-test | Warning: resource jaegers/with-sampling is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:21:31 | examples-with-sampling/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:21:31 | examples-with-sampling/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:21:31 | examples-with-sampling/2-smoke-test | job.batch/report-span created logger.go:42: 07:21:31 | examples-with-sampling/2-smoke-test | job.batch/check-span created logger.go:42: 07:21:43 | examples-with-sampling/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:21:43 | examples-with-sampling/3- | starting test step 3- logger.go:42: 07:21:43 | examples-with-sampling/3- | test step completed 3- logger.go:42: 07:21:43 | examples-with-sampling | examples-with-sampling events from ns kuttl-test-renewed-hare: logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:08 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-renewed-hare/cassandra-0 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:08 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:09 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.128.2.26/23] from ovn-kubernetes multus logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:09 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:13 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 3.924s (3.924s including waiting) kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:13 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:13 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:13 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-renewed-hare/cassandra-1 to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:13 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.131.0.42/23] from ovn-kubernetes multus logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:13 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:13 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:17 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 3.712s (3.712s including waiting) kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:17 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:17 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:22 +0000 UTC Normal Pod with-sampling-f9db774c6-mw6qq Binding Scheduled Successfully assigned kuttl-test-renewed-hare/with-sampling-f9db774c6-mw6qq to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:22 +0000 UTC Normal Pod with-sampling-f9db774c6-mw6qq AddedInterface Add eth0 [10.128.2.27/23] from ovn-kubernetes multus logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:22 +0000 UTC Normal Pod with-sampling-f9db774c6-mw6qq.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:22 +0000 UTC Normal Pod with-sampling-f9db774c6-mw6qq.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:22 +0000 UTC Normal Pod with-sampling-f9db774c6-mw6qq.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:22 +0000 UTC Normal Pod with-sampling-f9db774c6-mw6qq.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:22 +0000 UTC Normal ReplicaSet.apps with-sampling-f9db774c6 SuccessfulCreate Created pod: with-sampling-f9db774c6-mw6qq replicaset-controller logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:22 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-f9db774c6 to 1 deployment-controller logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:23 +0000 UTC Normal Pod with-sampling-f9db774c6-mw6qq.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:23 +0000 UTC Normal Pod with-sampling-f9db774c6-mw6qq.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:27 +0000 UTC Normal Pod with-sampling-f9db774c6-mw6qq.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:27 +0000 UTC Normal Pod with-sampling-f9db774c6-mw6qq.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:27 +0000 UTC Normal ReplicaSet.apps with-sampling-f9db774c6 SuccessfulDelete Deleted pod: with-sampling-f9db774c6-mw6qq replicaset-controller logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:27 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled down replica set with-sampling-f9db774c6 to 0 from 1 deployment-controller logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:28 +0000 UTC Normal Pod with-sampling-69f78d7946-78twd Binding Scheduled Successfully assigned kuttl-test-renewed-hare/with-sampling-69f78d7946-78twd to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:28 +0000 UTC Normal Pod with-sampling-69f78d7946-78twd AddedInterface Add eth0 [10.128.2.28/23] from ovn-kubernetes multus logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:28 +0000 UTC Normal Pod with-sampling-69f78d7946-78twd.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:28 +0000 UTC Normal ReplicaSet.apps with-sampling-69f78d7946 SuccessfulCreate Created pod: with-sampling-69f78d7946-78twd replicaset-controller logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:28 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-69f78d7946 to 1 deployment-controller logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:29 +0000 UTC Normal Pod with-sampling-69f78d7946-78twd.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:29 +0000 UTC Normal Pod with-sampling-69f78d7946-78twd.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:29 +0000 UTC Normal Pod with-sampling-69f78d7946-78twd.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:29 +0000 UTC Normal Pod with-sampling-69f78d7946-78twd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:29 +0000 UTC Normal Pod with-sampling-69f78d7946-78twd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:31 +0000 UTC Normal Pod check-span-fmzw9 Binding Scheduled Successfully assigned kuttl-test-renewed-hare/check-span-fmzw9 to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:31 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-fmzw9 job-controller logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:31 +0000 UTC Normal Pod report-span-gsxd2 Binding Scheduled Successfully assigned kuttl-test-renewed-hare/report-span-gsxd2 to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:31 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-gsxd2 job-controller logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:32 +0000 UTC Normal Pod check-span-fmzw9 AddedInterface Add eth0 [10.131.0.43/23] from ovn-kubernetes multus logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:32 +0000 UTC Normal Pod check-span-fmzw9.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:32 +0000 UTC Normal Pod check-span-fmzw9.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:32 +0000 UTC Normal Pod check-span-fmzw9.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:32 +0000 UTC Normal Pod report-span-gsxd2 AddedInterface Add eth0 [10.129.2.52/23] from ovn-kubernetes multus logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:32 +0000 UTC Normal Pod report-span-gsxd2.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:32 +0000 UTC Normal Pod report-span-gsxd2.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:32 +0000 UTC Normal Pod report-span-gsxd2.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:21:43 | examples-with-sampling | 2024-03-04 07:21:43 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:21:43 | examples-with-sampling | Deleting namespace: kuttl-test-renewed-hare === CONT kuttl/harness/examples-with-cassandra logger.go:42: 07:24:26 | examples-with-cassandra | Creating namespace: kuttl-test-glowing-duck logger.go:42: 07:24:26 | examples-with-cassandra/0-install | starting test step 0-install logger.go:42: 07:24:26 | examples-with-cassandra/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:24:26 | examples-with-cassandra/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:24:26 | examples-with-cassandra/0-install | >>>> Creating namespace kuttl-test-glowing-duck logger.go:42: 07:24:26 | examples-with-cassandra/0-install | kubectl create namespace kuttl-test-glowing-duck 2>&1 | grep -v "already exists" || true logger.go:42: 07:24:27 | examples-with-cassandra/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-glowing-duck 2>&1 | grep -v "already exists" || true logger.go:42: 07:24:27 | examples-with-cassandra/0-install | service/cassandra created logger.go:42: 07:24:27 | examples-with-cassandra/0-install | statefulset.apps/cassandra created logger.go:42: 07:24:27 | examples-with-cassandra/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:24:31 | examples-with-cassandra/0-install | test step completed 0-install logger.go:42: 07:24:31 | examples-with-cassandra/1-install | starting test step 1-install logger.go:42: 07:24:31 | examples-with-cassandra/1-install | Jaeger:kuttl-test-glowing-duck/with-cassandra created logger.go:42: 07:24:48 | examples-with-cassandra/1-install | test step completed 1-install logger.go:42: 07:24:48 | examples-with-cassandra/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:24:48 | examples-with-cassandra/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-cassandra /dev/null] logger.go:42: 07:24:49 | examples-with-cassandra/2-smoke-test | Warning: resource jaegers/with-cassandra is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:24:56 | examples-with-cassandra/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:24:56 | examples-with-cassandra/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:24:56 | examples-with-cassandra/2-smoke-test | job.batch/report-span created logger.go:42: 07:24:56 | examples-with-cassandra/2-smoke-test | job.batch/check-span created logger.go:42: 07:25:09 | examples-with-cassandra/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:25:09 | examples-with-cassandra | examples-with-cassandra events from ns kuttl-test-glowing-duck: logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:27 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-glowing-duck/cassandra-0 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:27 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:28 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.128.2.29/23] from ovn-kubernetes multus logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:28 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:28 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:28 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:28 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-glowing-duck/cassandra-1 to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:28 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:29 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.131.0.44/23] from ovn-kubernetes multus logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:29 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:29 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:29 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:34 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-kbxll Binding Scheduled Successfully assigned kuttl-test-glowing-duck/with-cassandra-cassandra-schema-job-kbxll to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:34 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job SuccessfulCreate Created pod: with-cassandra-cassandra-schema-job-kbxll job-controller logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:35 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-kbxll AddedInterface Add eth0 [10.129.2.53/23] from ovn-kubernetes multus logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:35 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-kbxll.spec.containers{with-cassandra-cassandra-schema-job} Pulling Pulling image "jaegertracing/jaeger-cassandra-schema:1.53.0" kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:39 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-kbxll.spec.containers{with-cassandra-cassandra-schema-job} Pulled Successfully pulled image "jaegertracing/jaeger-cassandra-schema:1.53.0" in 4.353s (4.353s including waiting) kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:39 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-kbxll.spec.containers{with-cassandra-cassandra-schema-job} Created Created container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:39 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-kbxll.spec.containers{with-cassandra-cassandra-schema-job} Started Started container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:45 +0000 UTC Normal Pod with-cassandra-5df7475f97-wd45b Binding Scheduled Successfully assigned kuttl-test-glowing-duck/with-cassandra-5df7475f97-wd45b to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:45 +0000 UTC Normal ReplicaSet.apps with-cassandra-5df7475f97 SuccessfulCreate Created pod: with-cassandra-5df7475f97-wd45b replicaset-controller logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:45 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job Completed Job completed job-controller logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:45 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-5df7475f97 to 1 deployment-controller logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:46 +0000 UTC Normal Pod with-cassandra-5df7475f97-wd45b AddedInterface Add eth0 [10.128.2.30/23] from ovn-kubernetes multus logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:46 +0000 UTC Normal Pod with-cassandra-5df7475f97-wd45b.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:46 +0000 UTC Normal Pod with-cassandra-5df7475f97-wd45b.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:46 +0000 UTC Normal Pod with-cassandra-5df7475f97-wd45b.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:46 +0000 UTC Normal Pod with-cassandra-5df7475f97-wd45b.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:46 +0000 UTC Normal Pod with-cassandra-5df7475f97-wd45b.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:46 +0000 UTC Normal Pod with-cassandra-5df7475f97-wd45b.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:52 +0000 UTC Normal Pod with-cassandra-5df7475f97-wd45b.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:52 +0000 UTC Normal Pod with-cassandra-5df7475f97-wd45b.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:52 +0000 UTC Normal ReplicaSet.apps with-cassandra-5df7475f97 SuccessfulDelete Deleted pod: with-cassandra-5df7475f97-wd45b replicaset-controller logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:52 +0000 UTC Normal Pod with-cassandra-5f79d98f7f-zvlrq Binding Scheduled Successfully assigned kuttl-test-glowing-duck/with-cassandra-5f79d98f7f-zvlrq to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:52 +0000 UTC Normal ReplicaSet.apps with-cassandra-5f79d98f7f SuccessfulCreate Created pod: with-cassandra-5f79d98f7f-zvlrq replicaset-controller logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:52 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled down replica set with-cassandra-5df7475f97 to 0 from 1 deployment-controller logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:52 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-5f79d98f7f to 1 deployment-controller logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:53 +0000 UTC Normal Pod with-cassandra-5f79d98f7f-zvlrq AddedInterface Add eth0 [10.128.2.31/23] from ovn-kubernetes multus logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:53 +0000 UTC Normal Pod with-cassandra-5f79d98f7f-zvlrq.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:53 +0000 UTC Normal Pod with-cassandra-5f79d98f7f-zvlrq.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:53 +0000 UTC Normal Pod with-cassandra-5f79d98f7f-zvlrq.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:53 +0000 UTC Normal Pod with-cassandra-5f79d98f7f-zvlrq.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:53 +0000 UTC Normal Pod with-cassandra-5f79d98f7f-zvlrq.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:53 +0000 UTC Normal Pod with-cassandra-5f79d98f7f-zvlrq.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:56 +0000 UTC Normal Pod check-span-29rbz Binding Scheduled Successfully assigned kuttl-test-glowing-duck/check-span-29rbz to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:56 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-29rbz job-controller logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:56 +0000 UTC Normal Pod report-span-jz7zt Binding Scheduled Successfully assigned kuttl-test-glowing-duck/report-span-jz7zt to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:56 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-jz7zt job-controller logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:57 +0000 UTC Normal Pod check-span-29rbz AddedInterface Add eth0 [10.131.0.45/23] from ovn-kubernetes multus logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:57 +0000 UTC Normal Pod check-span-29rbz.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:57 +0000 UTC Normal Pod check-span-29rbz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:57 +0000 UTC Normal Pod check-span-29rbz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:57 +0000 UTC Normal Pod report-span-jz7zt AddedInterface Add eth0 [10.129.2.54/23] from ovn-kubernetes multus logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:57 +0000 UTC Normal Pod report-span-jz7zt.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:57 +0000 UTC Normal Pod report-span-jz7zt.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:24:57 +0000 UTC Normal Pod report-span-jz7zt.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:25:09 | examples-with-cassandra | 2024-03-04 07:25:08 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:25:09 | examples-with-cassandra | Deleting namespace: kuttl-test-glowing-duck === CONT kuttl/harness/examples-with-badger logger.go:42: 07:25:21 | examples-with-badger | Creating namespace: kuttl-test-unique-newt logger.go:42: 07:25:21 | examples-with-badger/0-install | starting test step 0-install logger.go:42: 07:25:21 | examples-with-badger/0-install | Jaeger:kuttl-test-unique-newt/with-badger created logger.go:42: 07:25:27 | examples-with-badger/0-install | test step completed 0-install logger.go:42: 07:25:27 | examples-with-badger/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:25:27 | examples-with-badger/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger /dev/null] logger.go:42: 07:25:28 | examples-with-badger/1-smoke-test | Warning: resource jaegers/with-badger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:25:34 | examples-with-badger/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:25:35 | examples-with-badger/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:25:35 | examples-with-badger/1-smoke-test | job.batch/report-span created logger.go:42: 07:25:35 | examples-with-badger/1-smoke-test | job.batch/check-span created logger.go:42: 07:25:47 | examples-with-badger/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:25:47 | examples-with-badger | examples-with-badger events from ns kuttl-test-unique-newt: logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:24 +0000 UTC Normal Pod with-badger-768bc4777d-kp2pr Binding Scheduled Successfully assigned kuttl-test-unique-newt/with-badger-768bc4777d-kp2pr to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:24 +0000 UTC Normal ReplicaSet.apps with-badger-768bc4777d SuccessfulCreate Created pod: with-badger-768bc4777d-kp2pr replicaset-controller logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:24 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-768bc4777d to 1 deployment-controller logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:25 +0000 UTC Normal Pod with-badger-768bc4777d-kp2pr AddedInterface Add eth0 [10.128.2.32/23] from ovn-kubernetes multus logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:25 +0000 UTC Normal Pod with-badger-768bc4777d-kp2pr.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:25 +0000 UTC Normal Pod with-badger-768bc4777d-kp2pr.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:25 +0000 UTC Normal Pod with-badger-768bc4777d-kp2pr.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:25 +0000 UTC Normal Pod with-badger-768bc4777d-kp2pr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:25 +0000 UTC Normal Pod with-badger-768bc4777d-kp2pr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:25 +0000 UTC Normal Pod with-badger-768bc4777d-kp2pr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:29 +0000 UTC Normal Pod with-badger-768bc4777d-kp2pr.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:29 +0000 UTC Normal Pod with-badger-768bc4777d-kp2pr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:29 +0000 UTC Normal ReplicaSet.apps with-badger-768bc4777d SuccessfulDelete Deleted pod: with-badger-768bc4777d-kp2pr replicaset-controller logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:29 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled down replica set with-badger-768bc4777d to 0 from 1 deployment-controller logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:30 +0000 UTC Normal Pod with-badger-6d8b77747-5vfgx Binding Scheduled Successfully assigned kuttl-test-unique-newt/with-badger-6d8b77747-5vfgx to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:30 +0000 UTC Normal ReplicaSet.apps with-badger-6d8b77747 SuccessfulCreate Created pod: with-badger-6d8b77747-5vfgx replicaset-controller logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:30 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-6d8b77747 to 1 deployment-controller logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:31 +0000 UTC Normal Pod with-badger-6d8b77747-5vfgx AddedInterface Add eth0 [10.128.2.33/23] from ovn-kubernetes multus logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:31 +0000 UTC Normal Pod with-badger-6d8b77747-5vfgx.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:31 +0000 UTC Normal Pod with-badger-6d8b77747-5vfgx.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:31 +0000 UTC Normal Pod with-badger-6d8b77747-5vfgx.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:31 +0000 UTC Normal Pod with-badger-6d8b77747-5vfgx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:31 +0000 UTC Normal Pod with-badger-6d8b77747-5vfgx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:31 +0000 UTC Normal Pod with-badger-6d8b77747-5vfgx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:35 +0000 UTC Normal Pod check-span-6hhdj Binding Scheduled Successfully assigned kuttl-test-unique-newt/check-span-6hhdj to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:35 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-6hhdj job-controller logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:35 +0000 UTC Normal Pod report-span-dhtld Binding Scheduled Successfully assigned kuttl-test-unique-newt/report-span-dhtld to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:35 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-dhtld job-controller logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:36 +0000 UTC Normal Pod check-span-6hhdj AddedInterface Add eth0 [10.131.0.46/23] from ovn-kubernetes multus logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:36 +0000 UTC Normal Pod check-span-6hhdj.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:36 +0000 UTC Normal Pod check-span-6hhdj.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:36 +0000 UTC Normal Pod check-span-6hhdj.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:36 +0000 UTC Normal Pod report-span-dhtld AddedInterface Add eth0 [10.129.2.55/23] from ovn-kubernetes multus logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:36 +0000 UTC Normal Pod report-span-dhtld.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:36 +0000 UTC Normal Pod report-span-dhtld.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:36 +0000 UTC Normal Pod report-span-dhtld.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:25:47 | examples-with-badger | 2024-03-04 07:25:46 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:25:47 | examples-with-badger | Deleting namespace: kuttl-test-unique-newt === CONT kuttl/harness/examples-simplest logger.go:42: 07:25:59 | examples-simplest | Creating namespace: kuttl-test-sweet-quail logger.go:42: 07:25:59 | examples-simplest/0-install | starting test step 0-install logger.go:42: 07:25:59 | examples-simplest/0-install | Jaeger:kuttl-test-sweet-quail/simplest created logger.go:42: 07:26:06 | examples-simplest/0-install | test step completed 0-install logger.go:42: 07:26:06 | examples-simplest/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:26:06 | examples-simplest/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:26:07 | examples-simplest/1-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:26:14 | examples-simplest/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:26:14 | examples-simplest/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:26:14 | examples-simplest/1-smoke-test | job.batch/report-span created logger.go:42: 07:26:14 | examples-simplest/1-smoke-test | job.batch/check-span created logger.go:42: 07:26:26 | examples-simplest/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:26:26 | examples-simplest | examples-simplest events from ns kuttl-test-sweet-quail: logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:03 +0000 UTC Normal Pod simplest-6754856bb6-ksk75 Binding Scheduled Successfully assigned kuttl-test-sweet-quail/simplest-6754856bb6-ksk75 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:03 +0000 UTC Warning Pod simplest-6754856bb6-ksk75 FailedMount MountVolume.SetUp failed for volume "simplest-ui-oauth-proxy-tls" : secret "simplest-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:03 +0000 UTC Warning Pod simplest-6754856bb6-ksk75 FailedMount MountVolume.SetUp failed for volume "simplest-collector-tls-config-volume" : secret "simplest-collector-headless-tls" not found kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:03 +0000 UTC Normal ReplicaSet.apps simplest-6754856bb6 SuccessfulCreate Created pod: simplest-6754856bb6-ksk75 replicaset-controller logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:03 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-6754856bb6 to 1 deployment-controller logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:04 +0000 UTC Normal Pod simplest-6754856bb6-ksk75 AddedInterface Add eth0 [10.128.2.34/23] from ovn-kubernetes multus logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:04 +0000 UTC Normal Pod simplest-6754856bb6-ksk75.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:04 +0000 UTC Normal Pod simplest-6754856bb6-ksk75.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:04 +0000 UTC Normal Pod simplest-6754856bb6-ksk75.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:04 +0000 UTC Normal Pod simplest-6754856bb6-ksk75.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:04 +0000 UTC Normal Pod simplest-6754856bb6-ksk75.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:04 +0000 UTC Normal Pod simplest-6754856bb6-ksk75.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:10 +0000 UTC Normal Pod simplest-6754856bb6-ksk75.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:10 +0000 UTC Normal Pod simplest-6754856bb6-ksk75.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:10 +0000 UTC Normal ReplicaSet.apps simplest-6754856bb6 SuccessfulDelete Deleted pod: simplest-6754856bb6-ksk75 replicaset-controller logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:10 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-6754856bb6 to 0 from 1 deployment-controller logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:11 +0000 UTC Normal Pod simplest-cd7994958-sctv7 Binding Scheduled Successfully assigned kuttl-test-sweet-quail/simplest-cd7994958-sctv7 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:11 +0000 UTC Normal Pod simplest-cd7994958-sctv7 AddedInterface Add eth0 [10.128.2.35/23] from ovn-kubernetes multus logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:11 +0000 UTC Normal Pod simplest-cd7994958-sctv7.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:11 +0000 UTC Normal Pod simplest-cd7994958-sctv7.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:11 +0000 UTC Normal Pod simplest-cd7994958-sctv7.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:11 +0000 UTC Normal Pod simplest-cd7994958-sctv7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:11 +0000 UTC Normal Pod simplest-cd7994958-sctv7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:11 +0000 UTC Normal Pod simplest-cd7994958-sctv7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:11 +0000 UTC Normal ReplicaSet.apps simplest-cd7994958 SuccessfulCreate Created pod: simplest-cd7994958-sctv7 replicaset-controller logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:11 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-cd7994958 to 1 deployment-controller logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:14 +0000 UTC Normal Pod check-span-swjvk Binding Scheduled Successfully assigned kuttl-test-sweet-quail/check-span-swjvk to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:14 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-swjvk job-controller logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:14 +0000 UTC Normal Pod report-span-vj8sn Binding Scheduled Successfully assigned kuttl-test-sweet-quail/report-span-vj8sn to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:14 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-vj8sn job-controller logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:15 +0000 UTC Normal Pod check-span-swjvk AddedInterface Add eth0 [10.131.0.47/23] from ovn-kubernetes multus logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:15 +0000 UTC Normal Pod check-span-swjvk.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:15 +0000 UTC Normal Pod check-span-swjvk.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:15 +0000 UTC Normal Pod check-span-swjvk.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:15 +0000 UTC Normal Pod report-span-vj8sn AddedInterface Add eth0 [10.129.2.56/23] from ovn-kubernetes multus logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:15 +0000 UTC Normal Pod report-span-vj8sn.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:15 +0000 UTC Normal Pod report-span-vj8sn.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:15 +0000 UTC Normal Pod report-span-vj8sn.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:26:26 | examples-simplest | 2024-03-04 07:26:26 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:26:26 | examples-simplest | Deleting namespace: kuttl-test-sweet-quail === CONT kuttl/harness/examples-simple-prod-with-volumes logger.go:42: 07:26:39 | examples-simple-prod-with-volumes | Ignoring 03-check-volume.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:26:39 | examples-simple-prod-with-volumes | Creating namespace: kuttl-test-suitable-bulldog logger.go:42: 07:26:39 | examples-simple-prod-with-volumes/1-install | starting test step 1-install logger.go:42: 07:26:39 | examples-simple-prod-with-volumes/1-install | Jaeger:kuttl-test-suitable-bulldog/simple-prod created logger.go:42: 07:27:16 | examples-simple-prod-with-volumes/1-install | test step completed 1-install logger.go:42: 07:27:16 | examples-simple-prod-with-volumes/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:27:16 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:27:17 | examples-simple-prod-with-volumes/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:27:24 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:27:25 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:27:25 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/report-span created logger.go:42: 07:27:25 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/check-span created logger.go:42: 07:27:36 | examples-simple-prod-with-volumes/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:27:36 | examples-simple-prod-with-volumes/3-check-volume | starting test step 3-check-volume logger.go:42: 07:27:36 | examples-simple-prod-with-volumes/3-check-volume | running command: [sh -c kubectl exec $(kubectl get pods -n $NAMESPACE -l app=jaeger -l app.kubernetes.io/component=collector -o yaml | /tmp/jaeger-tests/bin/yq e '.items[0].metadata.name') -n $NAMESPACE -- ls /usr/share/elasticsearch/data] logger.go:42: 07:27:37 | examples-simple-prod-with-volumes/3-check-volume | test step completed 3-check-volume logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | examples-simple-prod-with-volumes events from ns kuttl-test-suitable-bulldog: logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:26:45 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestsuitablebulldogsimpleprod-1-5fc6cd988 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestsuitablebulldogsimpleprod-1-5fcvnss7 replicaset-controller logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:26:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsuitablebulldogsimpleprod-1-5fcvnss7 Binding Scheduled Successfully assigned kuttl-test-suitable-bulldog/elasticsearch-cdm-kuttltestsuitablebulldogsimpleprod-1-5fcvnss7 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:26:45 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestsuitablebulldogsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestsuitablebulldogsimpleprod-1-5fc6cd988 to 1 deployment-controller logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:26:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsuitablebulldogsimpleprod-1-5fcvnss7 AddedInterface Add eth0 [10.128.2.36/23] from ovn-kubernetes multus logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:26:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsuitablebulldogsimpleprod-1-5fcvnss7.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:26:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsuitablebulldogsimpleprod-1-5fcvnss7.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:26:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsuitablebulldogsimpleprod-1-5fcvnss7.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:26:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsuitablebulldogsimpleprod-1-5fcvnss7.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:26:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsuitablebulldogsimpleprod-1-5fcvnss7.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:26:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsuitablebulldogsimpleprod-1-5fcvnss7.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:26:56 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsuitablebulldogsimpleprod-1-5fcvnss7.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:01 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsuitablebulldogsimpleprod-1-5fcvnss7.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:12 +0000 UTC Normal Pod simple-prod-collector-d5bcf498-2cd56 Binding Scheduled Successfully assigned kuttl-test-suitable-bulldog/simple-prod-collector-d5bcf498-2cd56 to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:12 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-d5bcf498 SuccessfulCreate Created pod: simple-prod-collector-d5bcf498-2cd56 replicaset-controller logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:12 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-d5bcf498 to 1 deployment-controller logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:12 +0000 UTC Normal Pod simple-prod-query-6bf56dbbc-w5mgl Binding Scheduled Successfully assigned kuttl-test-suitable-bulldog/simple-prod-query-6bf56dbbc-w5mgl to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:12 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6bf56dbbc SuccessfulCreate Created pod: simple-prod-query-6bf56dbbc-w5mgl replicaset-controller logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:12 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6bf56dbbc to 1 deployment-controller logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:13 +0000 UTC Normal Pod simple-prod-collector-d5bcf498-2cd56 AddedInterface Add eth0 [10.129.2.57/23] from ovn-kubernetes multus logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:13 +0000 UTC Normal Pod simple-prod-collector-d5bcf498-2cd56.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" already present on machine kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:13 +0000 UTC Normal Pod simple-prod-collector-d5bcf498-2cd56.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:13 +0000 UTC Normal Pod simple-prod-collector-d5bcf498-2cd56.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:13 +0000 UTC Normal Pod simple-prod-query-6bf56dbbc-w5mgl AddedInterface Add eth0 [10.131.0.48/23] from ovn-kubernetes multus logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:13 +0000 UTC Normal Pod simple-prod-query-6bf56dbbc-w5mgl.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:13 +0000 UTC Normal Pod simple-prod-query-6bf56dbbc-w5mgl.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:13 +0000 UTC Normal Pod simple-prod-query-6bf56dbbc-w5mgl.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:13 +0000 UTC Normal Pod simple-prod-query-6bf56dbbc-w5mgl.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:13 +0000 UTC Normal Pod simple-prod-query-6bf56dbbc-w5mgl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:13 +0000 UTC Normal Pod simple-prod-query-6bf56dbbc-w5mgl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:13 +0000 UTC Normal Pod simple-prod-query-6bf56dbbc-w5mgl.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:13 +0000 UTC Normal Pod simple-prod-query-6bf56dbbc-w5mgl.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:13 +0000 UTC Normal Pod simple-prod-query-6bf56dbbc-w5mgl.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:21 +0000 UTC Normal Pod simple-prod-query-5fd7478456-h7v4m Binding Scheduled Successfully assigned kuttl-test-suitable-bulldog/simple-prod-query-5fd7478456-h7v4m to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:21 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5fd7478456 SuccessfulCreate Created pod: simple-prod-query-5fd7478456-h7v4m replicaset-controller logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:21 +0000 UTC Normal Pod simple-prod-query-6bf56dbbc-w5mgl.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:21 +0000 UTC Normal Pod simple-prod-query-6bf56dbbc-w5mgl.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:21 +0000 UTC Normal Pod simple-prod-query-6bf56dbbc-w5mgl.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:21 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6bf56dbbc SuccessfulDelete Deleted pod: simple-prod-query-6bf56dbbc-w5mgl replicaset-controller logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:21 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-6bf56dbbc to 0 from 1 deployment-controller logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:21 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-5fd7478456 to 1 deployment-controller logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:22 +0000 UTC Normal Pod simple-prod-query-5fd7478456-h7v4m AddedInterface Add eth0 [10.131.0.49/23] from ovn-kubernetes multus logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:22 +0000 UTC Normal Pod simple-prod-query-5fd7478456-h7v4m.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:22 +0000 UTC Normal Pod simple-prod-query-5fd7478456-h7v4m.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:22 +0000 UTC Normal Pod simple-prod-query-5fd7478456-h7v4m.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:22 +0000 UTC Normal Pod simple-prod-query-5fd7478456-h7v4m.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:22 +0000 UTC Normal Pod simple-prod-query-5fd7478456-h7v4m.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:22 +0000 UTC Normal Pod simple-prod-query-5fd7478456-h7v4m.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:22 +0000 UTC Normal Pod simple-prod-query-5fd7478456-h7v4m.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:22 +0000 UTC Normal Pod simple-prod-query-5fd7478456-h7v4m.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:22 +0000 UTC Normal Pod simple-prod-query-5fd7478456-h7v4m.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:25 +0000 UTC Normal Pod check-span-48wp6 Binding Scheduled Successfully assigned kuttl-test-suitable-bulldog/check-span-48wp6 to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:25 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-48wp6 job-controller logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:25 +0000 UTC Normal Pod report-span-h68m4 Binding Scheduled Successfully assigned kuttl-test-suitable-bulldog/report-span-h68m4 to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:25 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-h68m4 job-controller logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:26 +0000 UTC Normal Pod check-span-48wp6 AddedInterface Add eth0 [10.129.2.59/23] from ovn-kubernetes multus logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:26 +0000 UTC Normal Pod check-span-48wp6.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:26 +0000 UTC Normal Pod check-span-48wp6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:26 +0000 UTC Normal Pod check-span-48wp6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:26 +0000 UTC Normal Pod report-span-h68m4 AddedInterface Add eth0 [10.129.2.58/23] from ovn-kubernetes multus logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:26 +0000 UTC Normal Pod report-span-h68m4.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:26 +0000 UTC Normal Pod report-span-h68m4.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:26 +0000 UTC Normal Pod report-span-h68m4.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:27 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:27 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:27 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | 2024-03-04 07:27:36 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:27:37 | examples-simple-prod-with-volumes | Deleting namespace: kuttl-test-suitable-bulldog === CONT kuttl/harness/examples-business-application-injected-sidecar logger.go:42: 07:27:43 | examples-business-application-injected-sidecar | Creating namespace: kuttl-test-adapted-snake logger.go:42: 07:27:43 | examples-business-application-injected-sidecar/0-install | starting test step 0-install logger.go:42: 07:27:43 | examples-business-application-injected-sidecar/0-install | Deployment:kuttl-test-adapted-snake/myapp created logger.go:42: 07:27:43 | examples-business-application-injected-sidecar/0-install | test step completed 0-install logger.go:42: 07:27:43 | examples-business-application-injected-sidecar/1-install | starting test step 1-install logger.go:42: 07:27:43 | examples-business-application-injected-sidecar/1-install | Jaeger:kuttl-test-adapted-snake/simplest created logger.go:42: 07:27:54 | examples-business-application-injected-sidecar/1-install | test step completed 1-install logger.go:42: 07:27:54 | examples-business-application-injected-sidecar/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:27:54 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:27:55 | examples-business-application-injected-sidecar/2-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:28:02 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:28:02 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:28:02 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/report-span created logger.go:42: 07:28:02 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/check-span created logger.go:42: 07:28:14 | examples-business-application-injected-sidecar/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | examples-business-application-injected-sidecar events from ns kuttl-test-adapted-snake: logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:43 +0000 UTC Normal Pod myapp-679f79d5f8-tqj45 Binding Scheduled Successfully assigned kuttl-test-adapted-snake/myapp-679f79d5f8-tqj45 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:43 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulCreate Created pod: myapp-679f79d5f8-tqj45 replicaset-controller logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:43 +0000 UTC Normal Pod myapp-67b6797dc9-gghzm Binding Scheduled Successfully assigned kuttl-test-adapted-snake/myapp-67b6797dc9-gghzm to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:43 +0000 UTC Warning Pod myapp-67b6797dc9-gghzm FailedMount MountVolume.SetUp failed for volume "simplest-service-ca" : configmap "simplest-service-ca" not found kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:43 +0000 UTC Warning Pod myapp-67b6797dc9-gghzm FailedMount MountVolume.SetUp failed for volume "simplest-trusted-ca" : configmap "simplest-trusted-ca" not found kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:43 +0000 UTC Normal ReplicaSet.apps myapp-67b6797dc9 SuccessfulCreate Created pod: myapp-67b6797dc9-gghzm replicaset-controller logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:43 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-679f79d5f8 to 1 deployment-controller logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:43 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-67b6797dc9 to 1 deployment-controller logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:44 +0000 UTC Normal Pod myapp-679f79d5f8-tqj45 AddedInterface Add eth0 [10.128.2.37/23] from ovn-kubernetes multus logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:44 +0000 UTC Normal Pod myapp-679f79d5f8-tqj45.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:47 +0000 UTC Normal Pod myapp-679f79d5f8-tqj45.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.449s (3.449s including waiting) kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:47 +0000 UTC Normal Pod myapp-679f79d5f8-tqj45.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:47 +0000 UTC Normal Pod myapp-679f79d5f8-tqj45.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:51 +0000 UTC Normal Pod simplest-7d4dfd648b-4qk2r Binding Scheduled Successfully assigned kuttl-test-adapted-snake/simplest-7d4dfd648b-4qk2r to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:51 +0000 UTC Normal ReplicaSet.apps simplest-7d4dfd648b SuccessfulCreate Created pod: simplest-7d4dfd648b-4qk2r replicaset-controller logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:51 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-7d4dfd648b to 1 deployment-controller logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:52 +0000 UTC Warning Pod simplest-7d4dfd648b-4qk2r FailedMount MountVolume.SetUp failed for volume "simplest-collector-tls-config-volume" : secret "simplest-collector-headless-tls" not found kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:53 +0000 UTC Normal Pod simplest-7d4dfd648b-4qk2r AddedInterface Add eth0 [10.128.2.38/23] from ovn-kubernetes multus logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:53 +0000 UTC Normal Pod simplest-7d4dfd648b-4qk2r.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:53 +0000 UTC Normal Pod simplest-7d4dfd648b-4qk2r.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:53 +0000 UTC Normal Pod simplest-7d4dfd648b-4qk2r.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:53 +0000 UTC Normal Pod simplest-7d4dfd648b-4qk2r.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:53 +0000 UTC Normal Pod simplest-7d4dfd648b-4qk2r.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:53 +0000 UTC Normal Pod simplest-7d4dfd648b-4qk2r.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:54 +0000 UTC Warning Pod myapp-679f79d5f8-tqj45.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.128.2.37:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:57 +0000 UTC Normal Pod simplest-7d4dfd648b-4qk2r.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:57 +0000 UTC Normal Pod simplest-7d4dfd648b-4qk2r.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:57 +0000 UTC Normal ReplicaSet.apps simplest-7d4dfd648b SuccessfulDelete Deleted pod: simplest-7d4dfd648b-4qk2r replicaset-controller logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:57 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-7d4dfd648b to 0 from 1 deployment-controller logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:58 +0000 UTC Normal Pod simplest-6c8f75479-xlsf8 Binding Scheduled Successfully assigned kuttl-test-adapted-snake/simplest-6c8f75479-xlsf8 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:58 +0000 UTC Normal Pod simplest-6c8f75479-xlsf8 AddedInterface Add eth0 [10.128.2.39/23] from ovn-kubernetes multus logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:58 +0000 UTC Normal Pod simplest-6c8f75479-xlsf8.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:58 +0000 UTC Normal Pod simplest-6c8f75479-xlsf8.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:58 +0000 UTC Normal Pod simplest-6c8f75479-xlsf8.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:58 +0000 UTC Normal Pod simplest-6c8f75479-xlsf8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:58 +0000 UTC Normal ReplicaSet.apps simplest-6c8f75479 SuccessfulCreate Created pod: simplest-6c8f75479-xlsf8 replicaset-controller logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:58 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-6c8f75479 to 1 deployment-controller logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:59 +0000 UTC Normal Pod myapp-67b6797dc9-gghzm AddedInterface Add eth0 [10.129.2.60/23] from ovn-kubernetes multus logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:59 +0000 UTC Normal Pod myapp-67b6797dc9-gghzm.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:59 +0000 UTC Normal Pod simplest-6c8f75479-xlsf8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:27:59 +0000 UTC Normal Pod simplest-6c8f75479-xlsf8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:02 +0000 UTC Normal Pod check-span-4j2wp Binding Scheduled Successfully assigned kuttl-test-adapted-snake/check-span-4j2wp to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:02 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-4j2wp job-controller logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:02 +0000 UTC Normal Pod report-span-d27nx Binding Scheduled Successfully assigned kuttl-test-adapted-snake/report-span-d27nx to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:02 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-d27nx job-controller logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:03 +0000 UTC Normal Pod check-span-4j2wp AddedInterface Add eth0 [10.131.0.51/23] from ovn-kubernetes multus logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:03 +0000 UTC Normal Pod check-span-4j2wp.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:03 +0000 UTC Normal Pod check-span-4j2wp.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:03 +0000 UTC Normal Pod check-span-4j2wp.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:03 +0000 UTC Normal Pod myapp-67b6797dc9-gghzm.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.526s (3.526s including waiting) kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:03 +0000 UTC Normal Pod myapp-67b6797dc9-gghzm.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:03 +0000 UTC Normal Pod myapp-67b6797dc9-gghzm.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:03 +0000 UTC Normal Pod myapp-67b6797dc9-gghzm.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:03 +0000 UTC Normal Pod myapp-67b6797dc9-gghzm.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:03 +0000 UTC Normal Pod myapp-67b6797dc9-gghzm.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:03 +0000 UTC Normal Pod report-span-d27nx AddedInterface Add eth0 [10.131.0.50/23] from ovn-kubernetes multus logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:03 +0000 UTC Normal Pod report-span-d27nx.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:03 +0000 UTC Normal Pod report-span-d27nx.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:03 +0000 UTC Normal Pod report-span-d27nx.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:05 +0000 UTC Normal Pod myapp-679f79d5f8-tqj45.spec.containers{myapp} Killing Stopping container myapp kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:05 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulDelete Deleted pod: myapp-679f79d5f8-tqj45 replicaset-controller logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:05 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled down replica set myapp-679f79d5f8 to 0 from 1 deployment-controller logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:10 +0000 UTC Warning Pod myapp-67b6797dc9-gghzm.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.129.2.60:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | 2024-03-04 07:28:14 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:28:14 | examples-business-application-injected-sidecar | Deleting namespace: kuttl-test-adapted-snake === CONT kuttl/harness/examples-service-types logger.go:42: 07:28:25 | examples-service-types | Creating namespace: kuttl-test-useful-oyster logger.go:42: 07:28:26 | examples-service-types/0-install | starting test step 0-install logger.go:42: 07:28:26 | examples-service-types/0-install | Jaeger:kuttl-test-useful-oyster/service-types created logger.go:42: 07:28:32 | examples-service-types/0-install | test step completed 0-install logger.go:42: 07:28:32 | examples-service-types/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:28:32 | examples-service-types/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE service-types /dev/null] logger.go:42: 07:28:33 | examples-service-types/1-smoke-test | Warning: resource jaegers/service-types is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:28:40 | examples-service-types/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://service-types-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:28:40 | examples-service-types/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:28:41 | examples-service-types/1-smoke-test | job.batch/report-span created logger.go:42: 07:28:41 | examples-service-types/1-smoke-test | job.batch/check-span created logger.go:42: 07:28:52 | examples-service-types/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:28:52 | examples-service-types/2- | starting test step 2- logger.go:42: 07:28:52 | examples-service-types/2- | test step completed 2- logger.go:42: 07:28:52 | examples-service-types | examples-service-types events from ns kuttl-test-useful-oyster: logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:29 +0000 UTC Normal Pod service-types-5d476fb44c-pg85s Binding Scheduled Successfully assigned kuttl-test-useful-oyster/service-types-5d476fb44c-pg85s to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:29 +0000 UTC Normal ReplicaSet.apps service-types-5d476fb44c SuccessfulCreate Created pod: service-types-5d476fb44c-pg85s replicaset-controller logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:29 +0000 UTC Normal Service service-types-collector EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:29 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-5d476fb44c to 1 deployment-controller logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:30 +0000 UTC Normal Pod service-types-5d476fb44c-pg85s AddedInterface Add eth0 [10.128.2.40/23] from ovn-kubernetes multus logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:30 +0000 UTC Normal Pod service-types-5d476fb44c-pg85s.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:30 +0000 UTC Normal Pod service-types-5d476fb44c-pg85s.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:30 +0000 UTC Normal Pod service-types-5d476fb44c-pg85s.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:30 +0000 UTC Normal Pod service-types-5d476fb44c-pg85s.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:30 +0000 UTC Normal Pod service-types-5d476fb44c-pg85s.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:30 +0000 UTC Normal Pod service-types-5d476fb44c-pg85s.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:33 +0000 UTC Normal Service service-types-collector EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:33 +0000 UTC Normal Service service-types-query EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:36 +0000 UTC Normal Pod service-types-5d476fb44c-pg85s.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:36 +0000 UTC Normal Pod service-types-5d476fb44c-pg85s.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:36 +0000 UTC Normal ReplicaSet.apps service-types-5d476fb44c SuccessfulDelete Deleted pod: service-types-5d476fb44c-pg85s replicaset-controller logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:36 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled down replica set service-types-5d476fb44c to 0 from 1 deployment-controller logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:37 +0000 UTC Normal Pod service-types-5765944cd7-khfm9 Binding Scheduled Successfully assigned kuttl-test-useful-oyster/service-types-5765944cd7-khfm9 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:37 +0000 UTC Normal ReplicaSet.apps service-types-5765944cd7 SuccessfulCreate Created pod: service-types-5765944cd7-khfm9 replicaset-controller logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:37 +0000 UTC Normal Service service-types-query EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:37 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-5765944cd7 to 1 deployment-controller logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:38 +0000 UTC Normal Pod service-types-5765944cd7-khfm9 AddedInterface Add eth0 [10.128.2.41/23] from ovn-kubernetes multus logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:38 +0000 UTC Normal Pod service-types-5765944cd7-khfm9.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:38 +0000 UTC Normal Pod service-types-5765944cd7-khfm9.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:38 +0000 UTC Normal Pod service-types-5765944cd7-khfm9.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:38 +0000 UTC Normal Pod service-types-5765944cd7-khfm9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:38 +0000 UTC Normal Pod service-types-5765944cd7-khfm9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:38 +0000 UTC Normal Pod service-types-5765944cd7-khfm9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:41 +0000 UTC Normal Pod check-span-4vbqt Binding Scheduled Successfully assigned kuttl-test-useful-oyster/check-span-4vbqt to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:41 +0000 UTC Normal Pod check-span-4vbqt AddedInterface Add eth0 [10.131.0.52/23] from ovn-kubernetes multus logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:41 +0000 UTC Normal Pod check-span-4vbqt.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:41 +0000 UTC Normal Pod check-span-4vbqt.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:41 +0000 UTC Normal Pod check-span-4vbqt.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:41 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-4vbqt job-controller logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:41 +0000 UTC Normal Pod report-span-7kdt5 Binding Scheduled Successfully assigned kuttl-test-useful-oyster/report-span-7kdt5 to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:41 +0000 UTC Normal Pod report-span-7kdt5 AddedInterface Add eth0 [10.129.2.61/23] from ovn-kubernetes multus logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:41 +0000 UTC Normal Pod report-span-7kdt5.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:41 +0000 UTC Normal Pod report-span-7kdt5.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:41 +0000 UTC Normal Pod report-span-7kdt5.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:41 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-7kdt5 job-controller logger.go:42: 07:28:52 | examples-service-types | 2024-03-04 07:28:52 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:28:52 | examples-service-types | Deleting namespace: kuttl-test-useful-oyster === CONT kuttl/harness/examples-openshift-with-htpasswd logger.go:42: 07:29:24 | examples-openshift-with-htpasswd | Ignoring 00-install.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:29:24 | examples-openshift-with-htpasswd | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:29:24 | examples-openshift-with-htpasswd | Creating namespace: kuttl-test-faithful-wren logger.go:42: 07:29:24 | examples-openshift-with-htpasswd/0-install | starting test step 0-install logger.go:42: 07:29:24 | examples-openshift-with-htpasswd/0-install | Secret:kuttl-test-faithful-wren/htpasswd created logger.go:42: 07:29:24 | examples-openshift-with-htpasswd/0-install | test step completed 0-install logger.go:42: 07:29:24 | examples-openshift-with-htpasswd/1-install | starting test step 1-install logger.go:42: 07:29:24 | examples-openshift-with-htpasswd/1-install | Jaeger:kuttl-test-faithful-wren/with-htpasswd created logger.go:42: 07:29:29 | examples-openshift-with-htpasswd/1-install | test step completed 1-install logger.go:42: 07:29:29 | examples-openshift-with-htpasswd/2-check-unsecured | starting test step 2-check-unsecured logger.go:42: 07:29:29 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [./ensure-ingress-host.sh] logger.go:42: 07:29:29 | examples-openshift-with-htpasswd/2-check-unsecured | Checking the Ingress host value was populated logger.go:42: 07:29:29 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 0 logger.go:42: 07:29:29 | examples-openshift-with-htpasswd/2-check-unsecured | Hostname is with-htpasswd-kuttl-test-faithful-wren.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com logger.go:42: 07:29:29 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:29:29 | examples-openshift-with-htpasswd/2-check-unsecured | Checking an expected HTTP response logger.go:42: 07:29:29 | examples-openshift-with-htpasswd/2-check-unsecured | Running in OpenShift logger.go:42: 07:29:29 | examples-openshift-with-htpasswd/2-check-unsecured | Not using any secret logger.go:42: 07:29:29 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 1/30 the https://with-htpasswd-kuttl-test-faithful-wren.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:29:29 | examples-openshift-with-htpasswd/2-check-unsecured | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:29:29 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 2/30 the https://with-htpasswd-kuttl-test-faithful-wren.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:29:29 | examples-openshift-with-htpasswd/2-check-unsecured | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 07:29:39 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 3/30 the https://with-htpasswd-kuttl-test-faithful-wren.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:29:39 | examples-openshift-with-htpasswd/2-check-unsecured | curl response asserted properly logger.go:42: 07:29:39 | examples-openshift-with-htpasswd/2-check-unsecured | test step completed 2-check-unsecured logger.go:42: 07:29:39 | examples-openshift-with-htpasswd/3-check-unauthorized | starting test step 3-check-unauthorized logger.go:42: 07:29:39 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:29:39 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking the Ingress host value was populated logger.go:42: 07:29:39 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 0 logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/3-check-unauthorized | Hostname is with-htpasswd-kuttl-test-faithful-wren.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [sh -c JAEGER_USERNAME=wronguser JAEGER_PASSWORD=wrongpassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking an expected HTTP response logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/3-check-unauthorized | Running in OpenShift logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/3-check-unauthorized | Using Jaeger basic authentication logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 1/30 the https://with-htpasswd-kuttl-test-faithful-wren.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/3-check-unauthorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 2/30 the https://with-htpasswd-kuttl-test-faithful-wren.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/3-check-unauthorized | curl response asserted properly logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/3-check-unauthorized | test step completed 3-check-unauthorized logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/4-check-authorized | starting test step 4-check-authorized logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/4-check-authorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/4-check-authorized | Checking the Ingress host value was populated logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/4-check-authorized | Try number 0 logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/4-check-authorized | Hostname is with-htpasswd-kuttl-test-faithful-wren.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/4-check-authorized | running command: [sh -c JAEGER_USERNAME=awesomeuser JAEGER_PASSWORD=awesomepassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE with-htpasswd] logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/4-check-authorized | Checking an expected HTTP response logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/4-check-authorized | Running in OpenShift logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/4-check-authorized | Using Jaeger basic authentication logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/4-check-authorized | Try number 1/30 the https://with-htpasswd-kuttl-test-faithful-wren.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/4-check-authorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/4-check-authorized | Try number 2/30 the https://with-htpasswd-kuttl-test-faithful-wren.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/4-check-authorized | curl response asserted properly logger.go:42: 07:29:40 | examples-openshift-with-htpasswd/4-check-authorized | test step completed 4-check-authorized logger.go:42: 07:29:40 | examples-openshift-with-htpasswd | examples-openshift-with-htpasswd events from ns kuttl-test-faithful-wren: logger.go:42: 07:29:40 | examples-openshift-with-htpasswd | 2024-03-04 07:29:28 +0000 UTC Normal Pod with-htpasswd-85695bd697-xvk8n Binding Scheduled Successfully assigned kuttl-test-faithful-wren/with-htpasswd-85695bd697-xvk8n to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:29:40 | examples-openshift-with-htpasswd | 2024-03-04 07:29:28 +0000 UTC Normal Pod with-htpasswd-85695bd697-xvk8n AddedInterface Add eth0 [10.128.2.42/23] from ovn-kubernetes multus logger.go:42: 07:29:40 | examples-openshift-with-htpasswd | 2024-03-04 07:29:28 +0000 UTC Normal Pod with-htpasswd-85695bd697-xvk8n.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:29:40 | examples-openshift-with-htpasswd | 2024-03-04 07:29:28 +0000 UTC Normal Pod with-htpasswd-85695bd697-xvk8n.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:29:40 | examples-openshift-with-htpasswd | 2024-03-04 07:29:28 +0000 UTC Normal Pod with-htpasswd-85695bd697-xvk8n.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:29:40 | examples-openshift-with-htpasswd | 2024-03-04 07:29:28 +0000 UTC Normal Pod with-htpasswd-85695bd697-xvk8n.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:29:40 | examples-openshift-with-htpasswd | 2024-03-04 07:29:28 +0000 UTC Normal ReplicaSet.apps with-htpasswd-85695bd697 SuccessfulCreate Created pod: with-htpasswd-85695bd697-xvk8n replicaset-controller logger.go:42: 07:29:40 | examples-openshift-with-htpasswd | 2024-03-04 07:29:28 +0000 UTC Normal Deployment.apps with-htpasswd ScalingReplicaSet Scaled up replica set with-htpasswd-85695bd697 to 1 deployment-controller logger.go:42: 07:29:40 | examples-openshift-with-htpasswd | 2024-03-04 07:29:29 +0000 UTC Normal Pod with-htpasswd-85695bd697-xvk8n.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:29:40 | examples-openshift-with-htpasswd | 2024-03-04 07:29:29 +0000 UTC Normal Pod with-htpasswd-85695bd697-xvk8n.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:29:40 | examples-openshift-with-htpasswd | Deleting namespace: kuttl-test-faithful-wren === CONT kuttl/harness/examples-all-in-one-with-options logger.go:42: 07:29:46 | examples-all-in-one-with-options | Creating namespace: kuttl-test-charmed-asp logger.go:42: 07:29:46 | examples-all-in-one-with-options/0-install | starting test step 0-install logger.go:42: 07:29:46 | examples-all-in-one-with-options/0-install | Jaeger:kuttl-test-charmed-asp/my-jaeger created logger.go:42: 07:29:52 | examples-all-in-one-with-options/0-install | test step completed 0-install logger.go:42: 07:29:52 | examples-all-in-one-with-options/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:29:52 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:29:54 | examples-all-in-one-with-options/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:30:00 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443/jaeger MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:30:00 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:30:01 | examples-all-in-one-with-options/1-smoke-test | job.batch/report-span created logger.go:42: 07:30:01 | examples-all-in-one-with-options/1-smoke-test | job.batch/check-span created logger.go:42: 07:30:13 | examples-all-in-one-with-options/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:30:13 | examples-all-in-one-with-options | examples-all-in-one-with-options events from ns kuttl-test-charmed-asp: logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:50 +0000 UTC Normal Pod my-jaeger-7ff46ffb54-clhp7 Binding Scheduled Successfully assigned kuttl-test-charmed-asp/my-jaeger-7ff46ffb54-clhp7 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:50 +0000 UTC Normal ReplicaSet.apps my-jaeger-7ff46ffb54 SuccessfulCreate Created pod: my-jaeger-7ff46ffb54-clhp7 replicaset-controller logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:50 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-7ff46ffb54 to 1 deployment-controller logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:51 +0000 UTC Normal Pod my-jaeger-7ff46ffb54-clhp7 AddedInterface Add eth0 [10.128.2.43/23] from ovn-kubernetes multus logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:51 +0000 UTC Normal Pod my-jaeger-7ff46ffb54-clhp7.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:51 +0000 UTC Normal Pod my-jaeger-7ff46ffb54-clhp7.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:51 +0000 UTC Normal Pod my-jaeger-7ff46ffb54-clhp7.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:51 +0000 UTC Normal Pod my-jaeger-7ff46ffb54-clhp7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:51 +0000 UTC Normal Pod my-jaeger-7ff46ffb54-clhp7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:51 +0000 UTC Normal Pod my-jaeger-7ff46ffb54-clhp7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:55 +0000 UTC Normal Pod my-jaeger-7ff46ffb54-clhp7.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:55 +0000 UTC Normal Pod my-jaeger-7ff46ffb54-clhp7.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:55 +0000 UTC Normal ReplicaSet.apps my-jaeger-7ff46ffb54 SuccessfulDelete Deleted pod: my-jaeger-7ff46ffb54-clhp7 replicaset-controller logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:55 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-7ff46ffb54 to 0 from 1 deployment-controller logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:56 +0000 UTC Normal Pod my-jaeger-5bf765fb8-qwthp Binding Scheduled Successfully assigned kuttl-test-charmed-asp/my-jaeger-5bf765fb8-qwthp to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:56 +0000 UTC Normal ReplicaSet.apps my-jaeger-5bf765fb8 SuccessfulCreate Created pod: my-jaeger-5bf765fb8-qwthp replicaset-controller logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:56 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5bf765fb8 to 1 deployment-controller logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:57 +0000 UTC Normal Pod my-jaeger-5bf765fb8-qwthp AddedInterface Add eth0 [10.128.2.44/23] from ovn-kubernetes multus logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:57 +0000 UTC Normal Pod my-jaeger-5bf765fb8-qwthp.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:57 +0000 UTC Normal Pod my-jaeger-5bf765fb8-qwthp.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:57 +0000 UTC Normal Pod my-jaeger-5bf765fb8-qwthp.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:57 +0000 UTC Normal Pod my-jaeger-5bf765fb8-qwthp.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:57 +0000 UTC Normal Pod my-jaeger-5bf765fb8-qwthp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:29:57 +0000 UTC Normal Pod my-jaeger-5bf765fb8-qwthp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:30:01 +0000 UTC Normal Pod check-span-pxv56 Binding Scheduled Successfully assigned kuttl-test-charmed-asp/check-span-pxv56 to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:30:01 +0000 UTC Normal Pod check-span-pxv56 AddedInterface Add eth0 [10.131.0.53/23] from ovn-kubernetes multus logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:30:01 +0000 UTC Normal Pod check-span-pxv56.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:30:01 +0000 UTC Normal Pod check-span-pxv56.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:30:01 +0000 UTC Normal Pod check-span-pxv56.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:30:01 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-pxv56 job-controller logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:30:01 +0000 UTC Normal Pod report-span-k5r25 Binding Scheduled Successfully assigned kuttl-test-charmed-asp/report-span-k5r25 to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:30:01 +0000 UTC Normal Pod report-span-k5r25 AddedInterface Add eth0 [10.129.2.63/23] from ovn-kubernetes multus logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:30:01 +0000 UTC Normal Pod report-span-k5r25.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:30:01 +0000 UTC Normal Pod report-span-k5r25.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:30:01 +0000 UTC Normal Pod report-span-k5r25.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:30:01 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-k5r25 job-controller logger.go:42: 07:30:13 | examples-all-in-one-with-options | 2024-03-04 07:30:12 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:30:13 | examples-all-in-one-with-options | Deleting namespace: kuttl-test-charmed-asp === CONT kuttl/harness/examples-agent-with-priority-class logger.go:42: 07:30:20 | examples-agent-with-priority-class | Creating namespace: kuttl-test-viable-katydid logger.go:42: 07:30:20 | examples-agent-with-priority-class/0-install | starting test step 0-install logger.go:42: 07:30:20 | examples-agent-with-priority-class/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 07:30:20 | examples-agent-with-priority-class/0-install | ServiceAccount:kuttl-test-viable-katydid/jaeger-agent-daemonset created logger.go:42: 07:30:20 | examples-agent-with-priority-class/0-install | test step completed 0-install logger.go:42: 07:30:20 | examples-agent-with-priority-class/1-add-policy | starting test step 1-add-policy logger.go:42: 07:30:20 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 07:30:20 | examples-agent-with-priority-class/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 07:30:20 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 07:30:25 | examples-agent-with-priority-class/1-add-policy | test step completed 1-add-policy logger.go:42: 07:30:25 | examples-agent-with-priority-class/2-install | starting test step 2-install logger.go:42: 07:30:25 | examples-agent-with-priority-class/2-install | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 07:30:32 | examples-agent-with-priority-class/2-install | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:30:33 | examples-agent-with-priority-class/2-install | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:30:33 | examples-agent-with-priority-class/2-install | job.batch/report-span created logger.go:42: 07:30:33 | examples-agent-with-priority-class/2-install | job.batch/check-span created logger.go:42: 07:30:33 | examples-agent-with-priority-class/2-install | PriorityClass:/high-priority created logger.go:42: 07:30:33 | examples-agent-with-priority-class/2-install | Jaeger:kuttl-test-viable-katydid/agent-as-daemonset updated logger.go:42: 07:30:45 | examples-agent-with-priority-class/2-install | test step completed 2-install logger.go:42: 07:30:45 | examples-agent-with-priority-class | examples-agent-with-priority-class events from ns kuttl-test-viable-katydid: logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:30 +0000 UTC Normal Pod agent-as-daemonset-7bb55df484-j2bq6 Binding Scheduled Successfully assigned kuttl-test-viable-katydid/agent-as-daemonset-7bb55df484-j2bq6 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:30 +0000 UTC Normal Pod agent-as-daemonset-7bb55df484-j2bq6 AddedInterface Add eth0 [10.128.2.45/23] from ovn-kubernetes multus logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:30 +0000 UTC Normal Pod agent-as-daemonset-7bb55df484-j2bq6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:30 +0000 UTC Normal Pod agent-as-daemonset-7bb55df484-j2bq6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:30 +0000 UTC Normal Pod agent-as-daemonset-7bb55df484-j2bq6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:30 +0000 UTC Normal Pod agent-as-daemonset-7bb55df484-j2bq6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:30 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-7bb55df484 SuccessfulCreate Created pod: agent-as-daemonset-7bb55df484-j2bq6 replicaset-controller logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:30 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-7bb55df484 to 1 deployment-controller logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:31 +0000 UTC Normal Pod agent-as-daemonset-7bb55df484-j2bq6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:31 +0000 UTC Normal Pod agent-as-daemonset-7bb55df484-j2bq6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:33 +0000 UTC Normal Pod check-span-m4bx4 Binding Scheduled Successfully assigned kuttl-test-viable-katydid/check-span-m4bx4 to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:33 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-m4bx4 job-controller logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:33 +0000 UTC Normal Pod report-span-k6vzh Binding Scheduled Successfully assigned kuttl-test-viable-katydid/report-span-k6vzh to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:33 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-k6vzh job-controller logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:34 +0000 UTC Normal Pod check-span-m4bx4 AddedInterface Add eth0 [10.131.0.54/23] from ovn-kubernetes multus logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:34 +0000 UTC Normal Pod check-span-m4bx4.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:34 +0000 UTC Normal Pod check-span-m4bx4.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:34 +0000 UTC Normal Pod check-span-m4bx4.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:34 +0000 UTC Normal Pod report-span-k6vzh AddedInterface Add eth0 [10.129.2.64/23] from ovn-kubernetes multus logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:34 +0000 UTC Normal Pod report-span-k6vzh.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:34 +0000 UTC Normal Pod report-span-k6vzh.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:34 +0000 UTC Normal Pod report-span-k6vzh.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:37 +0000 UTC Warning DaemonSet.apps agent-as-daemonset-agent-daemonset FailedCreate Error creating: pods "agent-as-daemonset-agent-daemonset-" is forbidden: unable to validate against any security context constraint: [provider "anyuid": Forbidden: not usable by user or serviceaccount, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5775: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5778: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6831: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6832: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 14271: Host ports are not allowed to be used, provider "restricted": Forbidden: not usable by user or serviceaccount, provider "nonroot-v2": Forbidden: not usable by user or serviceaccount, provider "nonroot": Forbidden: not usable by user or serviceaccount, provider "hostmount-anyuid": Forbidden: not usable by user or serviceaccount, provider "elasticsearch-scc": Forbidden: not usable by user or serviceaccount, provider "machine-api-termination-handler": Forbidden: not usable by user or serviceaccount, provider "daemonset-with-hostport": Forbidden: not usable by user or serviceaccount, provider "hostnetwork-v2": Forbidden: not usable by user or serviceaccount, provider "hostnetwork": Forbidden: not usable by user or serviceaccount, provider "hostaccess": Forbidden: not usable by user or serviceaccount, provider "node-exporter": Forbidden: not usable by user or serviceaccount, provider "privileged": Forbidden: not usable by user or serviceaccount] daemonset-controller logger.go:42: 07:30:45 | examples-agent-with-priority-class | 2024-03-04 07:30:45 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:30:45 | examples-agent-with-priority-class | Deleting namespace: kuttl-test-viable-katydid === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (670.32s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.85s) --- PASS: kuttl/harness/examples-simple-prod (68.46s) --- PASS: kuttl/harness/examples-with-sampling (198.90s) --- PASS: kuttl/harness/examples-with-cassandra (54.29s) --- PASS: kuttl/harness/examples-with-badger (38.27s) --- PASS: kuttl/harness/examples-simplest (39.77s) --- PASS: kuttl/harness/examples-simple-prod-with-volumes (64.20s) --- PASS: kuttl/harness/examples-business-application-injected-sidecar (42.56s) --- PASS: kuttl/harness/examples-service-types (58.55s) --- PASS: kuttl/harness/examples-openshift-with-htpasswd (22.14s) --- PASS: kuttl/harness/examples-all-in-one-with-options (33.43s) --- PASS: kuttl/harness/examples-agent-with-priority-class (43.84s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name examples --report --output /logs/artifacts/examples.xml ./artifacts/kuttl-report.xml time="2024-03-04T07:31:04Z" level=debug msg="Setting a new name for the test suites" time="2024-03-04T07:31:04Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-03-04T07:31:04Z" level=debug msg="normalizing test case names" time="2024-03-04T07:31:04Z" level=debug msg="examples/artifacts -> examples_artifacts" time="2024-03-04T07:31:04Z" level=debug msg="examples/examples-simple-prod -> examples_examples_simple_prod" time="2024-03-04T07:31:04Z" level=debug msg="examples/examples-with-sampling -> examples_examples_with_sampling" time="2024-03-04T07:31:04Z" level=debug msg="examples/examples-with-cassandra -> examples_examples_with_cassandra" time="2024-03-04T07:31:04Z" level=debug msg="examples/examples-with-badger -> examples_examples_with_badger" time="2024-03-04T07:31:04Z" level=debug msg="examples/examples-simplest -> examples_examples_simplest" time="2024-03-04T07:31:04Z" level=debug msg="examples/examples-simple-prod-with-volumes -> examples_examples_simple_prod_with_volumes" time="2024-03-04T07:31:04Z" level=debug msg="examples/examples-business-application-injected-sidecar -> examples_examples_business_application_injected_sidecar" time="2024-03-04T07:31:04Z" level=debug msg="examples/examples-service-types -> examples_examples_service_types" time="2024-03-04T07:31:04Z" level=debug msg="examples/examples-openshift-with-htpasswd -> examples_examples_openshift_with_htpasswd" time="2024-03-04T07:31:04Z" level=debug msg="examples/examples-all-in-one-with-options -> examples_examples_all_in_one_with_options" time="2024-03-04T07:31:04Z" level=debug msg="examples/examples-agent-with-priority-class -> examples_examples_agent_with_priority_class" +---------------------------------------------------------+--------+ | NAME | RESULT | +---------------------------------------------------------+--------+ | examples_artifacts | passed | | examples_examples_simple_prod | passed | | examples_examples_with_sampling | passed | | examples_examples_with_cassandra | passed | | examples_examples_with_badger | passed | | examples_examples_simplest | passed | | examples_examples_simple_prod_with_volumes | passed | | examples_examples_business_application_injected_sidecar | passed | | examples_examples_service_types | passed | | examples_examples_openshift_with_htpasswd | passed | | examples_examples_all_in_one_with_options | passed | | examples_examples_agent_with_priority_class | passed | +---------------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh generate false true + '[' 3 -ne 3 ']' + test_suite_name=generate + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/generate.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-generate make[2]: Entering directory '/tmp/jaeger-tests' test -s /tmp/jaeger-tests/bin/operator-sdk || curl -sLo /tmp/jaeger-tests/bin/operator-sdk https://github.com/operator-framework/operator-sdk/releases/download/v1.32.0/operator-sdk_`go env GOOS`_`go env GOARCH` ./hack/install/install-golangci-lint.sh Installing golangci-lint golangci-lint 1.55.2 is installed already ./hack/install/install-goimports.sh Installing goimports Try 0... go install golang.org/x/tools/cmd/goimports@v0.1.12 >>>> Formatting code... ./.ci/format.sh >>>> Building... ./hack/install/install-dependencies.sh Installing go dependencies Try 0... go mod download GOOS= GOARCH= CGO_ENABLED=0 GO111MODULE=on go build -ldflags "-X "github.com/jaegertracing/jaeger-operator/pkg/version".version="1.54.0" -X "github.com/jaegertracing/jaeger-operator/pkg/version".buildDate=2024-03-04T07:31:05Z -X "github.com/jaegertracing/jaeger-operator/pkg/version".defaultJaeger="1.54.0"" -o "bin/jaeger-operator" main.go JAEGER_VERSION="1.54.0" ./tests/e2e/generate/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-02-29-062601 True False 43m Cluster version is 4.16.0-0.nightly-2024-02-29-062601' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-02-29-062601 True False 43m Cluster version is 4.16.0-0.nightly-2024-02-29-062601' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/generate/render.sh ++ export SUITE_DIR=./tests/e2e/generate ++ SUITE_DIR=./tests/e2e/generate ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/generate ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test generate 'This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 2 -ne 2 ']' + test_name=generate + message='This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/generate/_build + '[' _build '!=' _build ']' + rm -rf generate + warning 'generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed\e[0m' WAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running generate E2E tests' Running generate E2E tests + cd tests/e2e/generate/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2507913876 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:31:19 | artifacts | Creating namespace: kuttl-test-inviting-cub logger.go:42: 07:31:19 | artifacts | artifacts events from ns kuttl-test-inviting-cub: logger.go:42: 07:31:19 | artifacts | Deleting namespace: kuttl-test-inviting-cub === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (5.78s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.73s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name generate --report --output /logs/artifacts/generate.xml ./artifacts/kuttl-report.xml time="2024-03-04T07:31:25Z" level=debug msg="Setting a new name for the test suites" time="2024-03-04T07:31:25Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-03-04T07:31:25Z" level=debug msg="normalizing test case names" time="2024-03-04T07:31:25Z" level=debug msg="generate/artifacts -> generate_artifacts" +--------------------+--------+ | NAME | RESULT | +--------------------+--------+ | generate_artifacts | passed | +--------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh upgrade false true + '[' 3 -ne 3 ']' + test_suite_name=upgrade + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/upgrade.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-upgrade make[2]: Entering directory '/tmp/jaeger-tests' make docker JAEGER_VERSION=1.54.1 IMG="quay.io//jaeger-operator:next" make[3]: Entering directory '/tmp/jaeger-tests' [ ! -z "true" ] || docker build --build-arg=GOPROXY= --build-arg=VERSION="1.54.0" --build-arg=JAEGER_VERSION=1.54.1 --build-arg=TARGETARCH= --build-arg VERSION_DATE=2024-03-04T07:31:25Z --build-arg VERSION_PKG="github.com/jaegertracing/jaeger-operator/pkg/version" -t "quay.io//jaeger-operator:next" . make[3]: Leaving directory '/tmp/jaeger-tests' touch build-e2e-upgrade-image SKIP_ES_EXTERNAL=true IMG=quay.io//jaeger-operator:"1.54.0" JAEGER_OPERATOR_VERSION="1.54.0" JAEGER_VERSION="1.54.0" ./tests/e2e/upgrade/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-02-29-062601 True False 43m Cluster version is 4.16.0-0.nightly-2024-02-29-062601' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-02-29-062601 True False 43m Cluster version is 4.16.0-0.nightly-2024-02-29-062601' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ sort -V +++ tr ' ' '\n' +++ head -n 1 +++ echo 3.6.0 0.25.0 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/upgrade/render.sh ++ export SUITE_DIR=./tests/e2e/upgrade ++ SUITE_DIR=./tests/e2e/upgrade ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/upgrade ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + export JAEGER_NAME + '[' true = true ']' + skip_test upgrade 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade + warning 'upgrade: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade: Test not supported in OpenShift\e[0m' WAR: upgrade: Test not supported in OpenShift + '[' true = true ']' + skip_test upgrade-from-latest-release 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade-from-latest-release + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade-from-latest-release + warning 'upgrade-from-latest-release: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade-from-latest-release: Test not supported in OpenShift\e[0m' WAR: upgrade-from-latest-release: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running upgrade E2E tests' Running upgrade E2E tests + cd tests/e2e/upgrade/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2507913876 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:31:26 | artifacts | Creating namespace: kuttl-test-usable-mutt logger.go:42: 07:31:26 | artifacts | artifacts events from ns kuttl-test-usable-mutt: logger.go:42: 07:31:26 | artifacts | Deleting namespace: kuttl-test-usable-mutt === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (5.92s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.88s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name upgrade --report --output /logs/artifacts/upgrade.xml ./artifacts/kuttl-report.xml time="2024-03-04T07:31:32Z" level=debug msg="Setting a new name for the test suites" time="2024-03-04T07:31:32Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-03-04T07:31:32Z" level=debug msg="normalizing test case names" time="2024-03-04T07:31:32Z" level=debug msg="upgrade/artifacts -> upgrade_artifacts" +-------------------+--------+ | NAME | RESULT | +-------------------+--------+ | upgrade_artifacts | passed | +-------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh sidecar false true + '[' 3 -ne 3 ']' + test_suite_name=sidecar + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/sidecar.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-sidecar make[2]: Entering directory '/tmp/jaeger-tests' ./tests/e2e/sidecar/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-02-29-062601 True False 43m Cluster version is 4.16.0-0.nightly-2024-02-29-062601' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-02-29-062601 True False 43m Cluster version is 4.16.0-0.nightly-2024-02-29-062601' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/sidecar/render.sh ++ export SUITE_DIR=./tests/e2e/sidecar ++ SUITE_DIR=./tests/e2e/sidecar ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/sidecar ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + jaeger_service_name=order + start_test sidecar-deployment + '[' 1 -ne 1 ']' + test_name=sidecar-deployment + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-deployment' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-deployment\e[0m' Rendering files for test sidecar-deployment + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build + '[' _build '!=' _build ']' + mkdir -p sidecar-deployment + cd sidecar-deployment + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-namespace + '[' 1 -ne 1 ']' + test_name=sidecar-namespace + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-namespace' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-namespace\e[0m' Rendering files for test sidecar-namespace + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-deployment + '[' sidecar-deployment '!=' _build ']' + cd .. + mkdir -p sidecar-namespace + cd sidecar-namespace + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-skip-webhook + '[' 1 -ne 1 ']' + test_name=sidecar-skip-webhook + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-skip-webhook' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-skip-webhook\e[0m' Rendering files for test sidecar-skip-webhook + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-namespace + '[' sidecar-namespace '!=' _build ']' + cd .. + mkdir -p sidecar-skip-webhook + cd sidecar-skip-webhook + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running sidecar E2E tests' Running sidecar E2E tests + cd tests/e2e/sidecar/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2507913876 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/sidecar-deployment === PAUSE kuttl/harness/sidecar-deployment === RUN kuttl/harness/sidecar-namespace === PAUSE kuttl/harness/sidecar-namespace === RUN kuttl/harness/sidecar-skip-webhook === PAUSE kuttl/harness/sidecar-skip-webhook === CONT kuttl/harness/artifacts logger.go:42: 07:31:40 | artifacts | Creating namespace: kuttl-test-concrete-ocelot logger.go:42: 07:31:40 | artifacts | artifacts events from ns kuttl-test-concrete-ocelot: logger.go:42: 07:31:40 | artifacts | Deleting namespace: kuttl-test-concrete-ocelot === CONT kuttl/harness/sidecar-namespace logger.go:42: 07:31:46 | sidecar-namespace | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:31:46 | sidecar-namespace | Creating namespace: kuttl-test-special-hamster logger.go:42: 07:31:46 | sidecar-namespace/0-install | starting test step 0-install logger.go:42: 07:31:46 | sidecar-namespace/0-install | Jaeger:kuttl-test-special-hamster/agent-as-sidecar created logger.go:42: 07:31:52 | sidecar-namespace/0-install | test step completed 0-install logger.go:42: 07:31:52 | sidecar-namespace/1-install | starting test step 1-install logger.go:42: 07:31:52 | sidecar-namespace/1-install | Deployment:kuttl-test-special-hamster/vertx-create-span-sidecar created logger.go:42: 07:31:54 | sidecar-namespace/1-install | test step completed 1-install logger.go:42: 07:31:54 | sidecar-namespace/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:31:54 | sidecar-namespace/2-enable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="true"] logger.go:42: 07:31:54 | sidecar-namespace/2-enable-injection | namespace/kuttl-test-special-hamster annotated logger.go:42: 07:31:59 | sidecar-namespace/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:31:59 | sidecar-namespace/3-find-service | starting test step 3-find-service logger.go:42: 07:31:59 | sidecar-namespace/3-find-service | Job:kuttl-test-special-hamster/00-find-service created logger.go:42: 07:32:12 | sidecar-namespace/3-find-service | test step completed 3-find-service logger.go:42: 07:32:12 | sidecar-namespace/4-other-instance | starting test step 4-other-instance logger.go:42: 07:32:12 | sidecar-namespace/4-other-instance | Jaeger:kuttl-test-special-hamster/agent-as-sidecar2 created logger.go:42: 07:32:20 | sidecar-namespace/4-other-instance | test step completed 4-other-instance logger.go:42: 07:32:20 | sidecar-namespace/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:32:20 | sidecar-namespace/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:32:20 | sidecar-namespace/6-find-service | starting test step 6-find-service logger.go:42: 07:32:21 | sidecar-namespace/6-find-service | Job:kuttl-test-special-hamster/01-find-service created logger.go:42: 07:32:41 | sidecar-namespace/6-find-service | test step completed 6-find-service logger.go:42: 07:32:41 | sidecar-namespace/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:32:41 | sidecar-namespace/7-disable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="false"] logger.go:42: 07:32:41 | sidecar-namespace/7-disable-injection | namespace/kuttl-test-special-hamster annotated logger.go:42: 07:32:43 | sidecar-namespace/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:32:43 | sidecar-namespace | sidecar-namespace events from ns kuttl-test-special-hamster: logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:50 +0000 UTC Normal Pod agent-as-sidecar-59b977f684-fz9bm Binding Scheduled Successfully assigned kuttl-test-special-hamster/agent-as-sidecar-59b977f684-fz9bm to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:50 +0000 UTC Normal Pod agent-as-sidecar-59b977f684-fz9bm AddedInterface Add eth0 [10.128.2.46/23] from ovn-kubernetes multus logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:50 +0000 UTC Normal Pod agent-as-sidecar-59b977f684-fz9bm.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:50 +0000 UTC Normal Pod agent-as-sidecar-59b977f684-fz9bm.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:50 +0000 UTC Normal Pod agent-as-sidecar-59b977f684-fz9bm.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:50 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-59b977f684 SuccessfulCreate Created pod: agent-as-sidecar-59b977f684-fz9bm replicaset-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:50 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-59b977f684 to 1 deployment-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:52 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-4b9cz Binding Scheduled Successfully assigned kuttl-test-special-hamster/vertx-create-span-sidecar-84d458b68c-4b9cz to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:52 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-4b9cz replicaset-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:52 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:53 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-4b9cz AddedInterface Add eth0 [10.129.2.65/23] from ovn-kubernetes multus logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:53 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-4b9cz.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:53 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-4b9cz.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:53 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-4b9cz.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:54 +0000 UTC Normal Pod vertx-create-span-sidecar-56cddf994c-mbsbf Binding Scheduled Successfully assigned kuttl-test-special-hamster/vertx-create-span-sidecar-56cddf994c-mbsbf to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:54 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-56cddf994c SuccessfulCreate Created pod: vertx-create-span-sidecar-56cddf994c-mbsbf replicaset-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:54 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-56cddf994c to 1 deployment-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:55 +0000 UTC Normal Pod vertx-create-span-sidecar-56cddf994c-mbsbf AddedInterface Add eth0 [10.131.0.55/23] from ovn-kubernetes multus logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:55 +0000 UTC Normal Pod vertx-create-span-sidecar-56cddf994c-mbsbf.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:58 +0000 UTC Normal Pod vertx-create-span-sidecar-56cddf994c-mbsbf.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.353s (3.353s including waiting) kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:58 +0000 UTC Normal Pod vertx-create-span-sidecar-56cddf994c-mbsbf.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:58 +0000 UTC Normal Pod vertx-create-span-sidecar-56cddf994c-mbsbf.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:58 +0000 UTC Normal Pod vertx-create-span-sidecar-56cddf994c-mbsbf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:58 +0000 UTC Normal Pod vertx-create-span-sidecar-56cddf994c-mbsbf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:58 +0000 UTC Normal Pod vertx-create-span-sidecar-56cddf994c-mbsbf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:59 +0000 UTC Normal Pod 00-find-service-kb5tj Binding Scheduled Successfully assigned kuttl-test-special-hamster/00-find-service-kb5tj to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:31:59 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-kb5tj job-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:00 +0000 UTC Normal Pod 00-find-service-kb5tj AddedInterface Add eth0 [10.128.2.47/23] from ovn-kubernetes multus logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:00 +0000 UTC Normal Pod 00-find-service-kb5tj.spec.containers{asserts-container} Pulling Pulling image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:01 +0000 UTC Normal Pod 00-find-service-kb5tj.spec.containers{asserts-container} Pulled Successfully pulled image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" in 1.421s (1.421s including waiting) kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:01 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-4b9cz.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.65:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:01 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-4b9cz.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.65:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:02 +0000 UTC Normal Pod 00-find-service-kb5tj.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:02 +0000 UTC Normal Pod 00-find-service-kb5tj.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:03 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-4b9cz.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:03 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-4b9cz.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.65:8080/": read tcp 10.129.2.2:41594->10.129.2.65:8080: read: connection reset by peer kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:03 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-4b9cz.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.65:8080/": dial tcp 10.129.2.65:8080: connect: connection refused kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:06 +0000 UTC Warning Pod vertx-create-span-sidecar-56cddf994c-mbsbf.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.55:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:06 +0000 UTC Warning Pod vertx-create-span-sidecar-56cddf994c-mbsbf.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.55:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:08 +0000 UTC Normal Pod vertx-create-span-sidecar-56cddf994c-mbsbf.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:08 +0000 UTC Warning Pod vertx-create-span-sidecar-56cddf994c-mbsbf.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.55:8080/": read tcp 10.131.0.2:41070->10.131.0.55:8080: read: connection reset by peer kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:08 +0000 UTC Warning Pod vertx-create-span-sidecar-56cddf994c-mbsbf.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.55:8080/": dial tcp 10.131.0.55:8080: connect: connection refused kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:08 +0000 UTC Normal Pod vertx-create-span-sidecar-56cddf994c-mbsbf.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:12 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:13 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-4b9cz.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.65:8080/": read tcp 10.129.2.2:52820->10.129.2.65:8080: read: connection reset by peer kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:18 +0000 UTC Normal Pod agent-as-sidecar2-74dd8cc45d-2k2ht Binding Scheduled Successfully assigned kuttl-test-special-hamster/agent-as-sidecar2-74dd8cc45d-2k2ht to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:18 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-74dd8cc45d SuccessfulCreate Created pod: agent-as-sidecar2-74dd8cc45d-2k2ht replicaset-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:18 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-74dd8cc45d to 1 deployment-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:19 +0000 UTC Normal Pod agent-as-sidecar2-74dd8cc45d-2k2ht AddedInterface Add eth0 [10.128.2.48/23] from ovn-kubernetes multus logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:19 +0000 UTC Normal Pod agent-as-sidecar2-74dd8cc45d-2k2ht.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:19 +0000 UTC Normal Pod agent-as-sidecar2-74dd8cc45d-2k2ht.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:19 +0000 UTC Normal Pod agent-as-sidecar2-74dd8cc45d-2k2ht.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:21 +0000 UTC Normal Pod 01-find-service-pb5b7 Binding Scheduled Successfully assigned kuttl-test-special-hamster/01-find-service-pb5b7 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:21 +0000 UTC Normal Pod 01-find-service-pb5b7 AddedInterface Add eth0 [10.128.2.49/23] from ovn-kubernetes multus logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:21 +0000 UTC Normal Pod 01-find-service-pb5b7.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:21 +0000 UTC Normal Pod 01-find-service-pb5b7.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:21 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-pb5b7 job-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:21 +0000 UTC Normal Pod agent-as-sidecar-59b977f684-fz9bm.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:22 +0000 UTC Normal Pod 01-find-service-pb5b7.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:24 +0000 UTC Normal Pod vertx-create-span-sidecar-77679ddbd5-8lmf5 Binding Scheduled Successfully assigned kuttl-test-special-hamster/vertx-create-span-sidecar-77679ddbd5-8lmf5 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:24 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-77679ddbd5 SuccessfulCreate Created pod: vertx-create-span-sidecar-77679ddbd5-8lmf5 replicaset-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:24 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-4b9cz replicaset-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:24 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:24 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-77679ddbd5 to 1 from 0 deployment-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:25 +0000 UTC Normal Pod vertx-create-span-sidecar-77679ddbd5-8lmf5 AddedInterface Add eth0 [10.128.2.50/23] from ovn-kubernetes multus logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:25 +0000 UTC Normal Pod vertx-create-span-sidecar-77679ddbd5-8lmf5.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:25 +0000 UTC Normal Pod vertx-create-span-sidecar-77679ddbd5-8lmf5.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:25 +0000 UTC Normal Pod vertx-create-span-sidecar-77679ddbd5-8lmf5.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:25 +0000 UTC Normal Pod vertx-create-span-sidecar-77679ddbd5-8lmf5.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:25 +0000 UTC Normal Pod vertx-create-span-sidecar-77679ddbd5-8lmf5.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:25 +0000 UTC Normal Pod vertx-create-span-sidecar-77679ddbd5-8lmf5.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:33 +0000 UTC Warning Pod vertx-create-span-sidecar-77679ddbd5-8lmf5.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.50:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:33 +0000 UTC Warning Pod vertx-create-span-sidecar-77679ddbd5-8lmf5.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.50:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:35 +0000 UTC Normal Pod vertx-create-span-sidecar-77679ddbd5-8lmf5.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:36 +0000 UTC Warning Pod vertx-create-span-sidecar-77679ddbd5-8lmf5.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.50:8080/": read tcp 10.128.2.2:50170->10.128.2.50:8080: read: connection reset by peer kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:36 +0000 UTC Warning Pod vertx-create-span-sidecar-77679ddbd5-8lmf5.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.50:8080/": dial tcp 10.128.2.50:8080: connect: connection refused kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:40 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:41 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-56cddf994c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-56cddf994c-mbsbf replicaset-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:41 +0000 UTC Normal Pod vertx-create-span-sidecar-5f8d58888-s4fsp Binding Scheduled Successfully assigned kuttl-test-special-hamster/vertx-create-span-sidecar-5f8d58888-s4fsp to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:41 +0000 UTC Normal Pod vertx-create-span-sidecar-5f8d58888-s4fsp AddedInterface Add eth0 [10.129.2.66/23] from ovn-kubernetes multus logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:41 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-5f8d58888 SuccessfulCreate Created pod: vertx-create-span-sidecar-5f8d58888-s4fsp replicaset-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:41 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-56cddf994c to 0 from 1 deployment-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:41 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-5f8d58888 to 1 from 0 deployment-controller logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:42 +0000 UTC Normal Pod vertx-create-span-sidecar-5f8d58888-s4fsp.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:42 +0000 UTC Normal Pod vertx-create-span-sidecar-5f8d58888-s4fsp.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:32:43 | sidecar-namespace | 2024-03-04 07:32:42 +0000 UTC Normal Pod vertx-create-span-sidecar-5f8d58888-s4fsp.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:32:43 | sidecar-namespace | Deleting namespace: kuttl-test-special-hamster === CONT kuttl/harness/sidecar-skip-webhook logger.go:42: 07:32:49 | sidecar-skip-webhook | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:32:49 | sidecar-skip-webhook | Creating namespace: kuttl-test-guiding-reptile logger.go:42: 07:32:49 | sidecar-skip-webhook/0-install | starting test step 0-install logger.go:42: 07:32:49 | sidecar-skip-webhook/0-install | Jaeger:kuttl-test-guiding-reptile/agent-as-sidecar created logger.go:42: 07:32:55 | sidecar-skip-webhook/0-install | test step completed 0-install logger.go:42: 07:32:55 | sidecar-skip-webhook/1-install | starting test step 1-install logger.go:42: 07:32:56 | sidecar-skip-webhook/1-install | Deployment:kuttl-test-guiding-reptile/vertx-create-span-sidecar created logger.go:42: 07:32:58 | sidecar-skip-webhook/1-install | test step completed 1-install logger.go:42: 07:32:58 | sidecar-skip-webhook/2-add-anotation-and-label | starting test step 2-add-anotation-and-label logger.go:42: 07:32:58 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name=jaeger-operator --namespace kuttl-test-guiding-reptile] logger.go:42: 07:32:58 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar labeled logger.go:42: 07:32:58 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-guiding-reptile] logger.go:42: 07:32:58 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:32:58 | sidecar-skip-webhook/2-add-anotation-and-label | test step completed 2-add-anotation-and-label logger.go:42: 07:32:58 | sidecar-skip-webhook/3-remove-label | starting test step 3-remove-label logger.go:42: 07:32:58 | sidecar-skip-webhook/3-remove-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name- --namespace kuttl-test-guiding-reptile] logger.go:42: 07:32:58 | sidecar-skip-webhook/3-remove-label | deployment.apps/vertx-create-span-sidecar unlabeled logger.go:42: 07:33:00 | sidecar-skip-webhook/3-remove-label | test step completed 3-remove-label logger.go:42: 07:33:00 | sidecar-skip-webhook | sidecar-skip-webhook events from ns kuttl-test-guiding-reptile: logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:53 +0000 UTC Normal Pod agent-as-sidecar-6cdb8cb95-d8fc6 Binding Scheduled Successfully assigned kuttl-test-guiding-reptile/agent-as-sidecar-6cdb8cb95-d8fc6 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:53 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-6cdb8cb95 SuccessfulCreate Created pod: agent-as-sidecar-6cdb8cb95-d8fc6 replicaset-controller logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:53 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-6cdb8cb95 to 1 deployment-controller logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:54 +0000 UTC Normal Pod agent-as-sidecar-6cdb8cb95-d8fc6 AddedInterface Add eth0 [10.128.2.51/23] from ovn-kubernetes multus logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:54 +0000 UTC Normal Pod agent-as-sidecar-6cdb8cb95-d8fc6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:54 +0000 UTC Normal Pod agent-as-sidecar-6cdb8cb95-d8fc6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:54 +0000 UTC Normal Pod agent-as-sidecar-6cdb8cb95-d8fc6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:56 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-9jktf Binding Scheduled Successfully assigned kuttl-test-guiding-reptile/vertx-create-span-sidecar-84d458b68c-9jktf to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:56 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-9jktf AddedInterface Add eth0 [10.129.2.67/23] from ovn-kubernetes multus logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:56 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-9jktf.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:56 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-9jktf.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:56 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-9jktf.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:56 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-9jktf replicaset-controller logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:56 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:58 +0000 UTC Normal Pod vertx-create-span-sidecar-7bc7bfd54b-6cnm6 Binding Scheduled Successfully assigned kuttl-test-guiding-reptile/vertx-create-span-sidecar-7bc7bfd54b-6cnm6 to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:58 +0000 UTC Normal Pod vertx-create-span-sidecar-7bc7bfd54b-6cnm6 AddedInterface Add eth0 [10.131.0.56/23] from ovn-kubernetes multus logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:58 +0000 UTC Normal Pod vertx-create-span-sidecar-7bc7bfd54b-6cnm6.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:58 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-7bc7bfd54b SuccessfulCreate Created pod: vertx-create-span-sidecar-7bc7bfd54b-6cnm6 replicaset-controller logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:58 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-7bc7bfd54b to 1 deployment-controller logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:59 +0000 UTC Normal Pod vertx-create-span-sidecar-7bc7bfd54b-6cnm6.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:59 +0000 UTC Normal Pod vertx-create-span-sidecar-7bc7bfd54b-6cnm6.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:59 +0000 UTC Normal Pod vertx-create-span-sidecar-7bc7bfd54b-6cnm6.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:59 +0000 UTC Normal Pod vertx-create-span-sidecar-7bc7bfd54b-6cnm6.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:33:00 | sidecar-skip-webhook | 2024-03-04 07:32:59 +0000 UTC Normal Pod vertx-create-span-sidecar-7bc7bfd54b-6cnm6.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:33:00 | sidecar-skip-webhook | Deleting namespace: kuttl-test-guiding-reptile === CONT kuttl/harness/sidecar-deployment logger.go:42: 07:33:06 | sidecar-deployment | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:33:06 | sidecar-deployment | Creating namespace: kuttl-test-workable-kodiak logger.go:42: 07:33:06 | sidecar-deployment/0-install | starting test step 0-install logger.go:42: 07:33:06 | sidecar-deployment/0-install | Jaeger:kuttl-test-workable-kodiak/agent-as-sidecar created logger.go:42: 07:33:12 | sidecar-deployment/0-install | test step completed 0-install logger.go:42: 07:33:12 | sidecar-deployment/1-install | starting test step 1-install logger.go:42: 07:33:12 | sidecar-deployment/1-install | Deployment:kuttl-test-workable-kodiak/vertx-create-span-sidecar created logger.go:42: 07:33:13 | sidecar-deployment/1-install | test step completed 1-install logger.go:42: 07:33:13 | sidecar-deployment/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:33:13 | sidecar-deployment/2-enable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-workable-kodiak] logger.go:42: 07:33:13 | sidecar-deployment/2-enable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:33:15 | sidecar-deployment/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:33:15 | sidecar-deployment/3-find-service | starting test step 3-find-service logger.go:42: 07:33:15 | sidecar-deployment/3-find-service | Job:kuttl-test-workable-kodiak/00-find-service created logger.go:42: 07:33:27 | sidecar-deployment/3-find-service | test step completed 3-find-service logger.go:42: 07:33:27 | sidecar-deployment/4-other-instance | starting test step 4-other-instance logger.go:42: 07:33:27 | sidecar-deployment/4-other-instance | Jaeger:kuttl-test-workable-kodiak/agent-as-sidecar2 created logger.go:42: 07:33:35 | sidecar-deployment/4-other-instance | test step completed 4-other-instance logger.go:42: 07:33:35 | sidecar-deployment/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:33:36 | sidecar-deployment/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:33:36 | sidecar-deployment/6-find-service | starting test step 6-find-service logger.go:42: 07:33:37 | sidecar-deployment/6-find-service | Job:kuttl-test-workable-kodiak/01-find-service created logger.go:42: 07:33:56 | sidecar-deployment/6-find-service | test step completed 6-find-service logger.go:42: 07:33:56 | sidecar-deployment/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:33:56 | sidecar-deployment/7-disable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=false --namespace kuttl-test-workable-kodiak] logger.go:42: 07:33:56 | sidecar-deployment/7-disable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:33:58 | sidecar-deployment/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:33:58 | sidecar-deployment | sidecar-deployment events from ns kuttl-test-workable-kodiak: logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:10 +0000 UTC Normal Pod agent-as-sidecar-86b878cb8c-p95kv Binding Scheduled Successfully assigned kuttl-test-workable-kodiak/agent-as-sidecar-86b878cb8c-p95kv to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:10 +0000 UTC Normal Pod agent-as-sidecar-86b878cb8c-p95kv AddedInterface Add eth0 [10.128.2.52/23] from ovn-kubernetes multus logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:10 +0000 UTC Normal Pod agent-as-sidecar-86b878cb8c-p95kv.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:10 +0000 UTC Normal Pod agent-as-sidecar-86b878cb8c-p95kv.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:10 +0000 UTC Normal Pod agent-as-sidecar-86b878cb8c-p95kv.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:10 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-86b878cb8c SuccessfulCreate Created pod: agent-as-sidecar-86b878cb8c-p95kv replicaset-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:10 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-86b878cb8c to 1 deployment-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:12 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-jpj6f Binding Scheduled Successfully assigned kuttl-test-workable-kodiak/vertx-create-span-sidecar-84d458b68c-jpj6f to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:12 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-jpj6f replicaset-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:12 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:13 +0000 UTC Normal Pod vertx-create-span-sidecar-68cffc55cd-dml9d Binding Scheduled Successfully assigned kuttl-test-workable-kodiak/vertx-create-span-sidecar-68cffc55cd-dml9d to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:13 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-68cffc55cd SuccessfulCreate Created pod: vertx-create-span-sidecar-68cffc55cd-dml9d replicaset-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:13 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-jpj6f AddedInterface Add eth0 [10.129.2.68/23] from ovn-kubernetes multus logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:13 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-jpj6f.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:13 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-jpj6f.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:13 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-jpj6f.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:13 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-68cffc55cd to 1 deployment-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:14 +0000 UTC Normal Pod vertx-create-span-sidecar-68cffc55cd-dml9d AddedInterface Add eth0 [10.128.2.53/23] from ovn-kubernetes multus logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:14 +0000 UTC Normal Pod vertx-create-span-sidecar-68cffc55cd-dml9d.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:14 +0000 UTC Normal Pod vertx-create-span-sidecar-68cffc55cd-dml9d.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:14 +0000 UTC Normal Pod vertx-create-span-sidecar-68cffc55cd-dml9d.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:14 +0000 UTC Normal Pod vertx-create-span-sidecar-68cffc55cd-dml9d.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:14 +0000 UTC Normal Pod vertx-create-span-sidecar-68cffc55cd-dml9d.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:14 +0000 UTC Normal Pod vertx-create-span-sidecar-68cffc55cd-dml9d.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:15 +0000 UTC Normal Pod 00-find-service-j8gsl Binding Scheduled Successfully assigned kuttl-test-workable-kodiak/00-find-service-j8gsl to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:15 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-j8gsl job-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:16 +0000 UTC Normal Pod 00-find-service-j8gsl AddedInterface Add eth0 [10.131.0.57/23] from ovn-kubernetes multus logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:16 +0000 UTC Normal Pod 00-find-service-j8gsl.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:16 +0000 UTC Normal Pod 00-find-service-j8gsl.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:16 +0000 UTC Normal Pod 00-find-service-j8gsl.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:21 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-jpj6f.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.68:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:21 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-jpj6f.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.68:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:22 +0000 UTC Warning Pod vertx-create-span-sidecar-68cffc55cd-dml9d.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.53:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:22 +0000 UTC Warning Pod vertx-create-span-sidecar-68cffc55cd-dml9d.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.53:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:24 +0000 UTC Normal Pod vertx-create-span-sidecar-68cffc55cd-dml9d.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:24 +0000 UTC Warning Pod vertx-create-span-sidecar-68cffc55cd-dml9d.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.53:8080/": read tcp 10.128.2.2:58692->10.128.2.53:8080: read: connection reset by peer kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:24 +0000 UTC Warning Pod vertx-create-span-sidecar-68cffc55cd-dml9d.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.53:8080/": dial tcp 10.128.2.53:8080: connect: connection refused kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:24 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-jpj6f.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:24 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-jpj6f.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.68:8080/": read tcp 10.129.2.2:56444->10.129.2.68:8080: read: connection reset by peer kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:24 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-jpj6f.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.68:8080/": dial tcp 10.129.2.68:8080: connect: connection refused kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:26 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:31 +0000 UTC Normal Pod agent-as-sidecar2-5f68587c4c-6pd4k Binding Scheduled Successfully assigned kuttl-test-workable-kodiak/agent-as-sidecar2-5f68587c4c-6pd4k to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:31 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-5f68587c4c SuccessfulCreate Created pod: agent-as-sidecar2-5f68587c4c-6pd4k replicaset-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:31 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-5f68587c4c to 1 deployment-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:32 +0000 UTC Normal Pod agent-as-sidecar2-5f68587c4c-6pd4k AddedInterface Add eth0 [10.131.0.58/23] from ovn-kubernetes multus logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:32 +0000 UTC Normal Pod agent-as-sidecar2-5f68587c4c-6pd4k.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:34 +0000 UTC Normal Pod agent-as-sidecar2-5f68587c4c-6pd4k.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" in 1.598s (1.598s including waiting) kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:34 +0000 UTC Normal Pod agent-as-sidecar2-5f68587c4c-6pd4k.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:34 +0000 UTC Normal Pod agent-as-sidecar2-5f68587c4c-6pd4k.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:35 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-jpj6f.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.68:8080/": read tcp 10.129.2.2:58600->10.129.2.68:8080: read: connection reset by peer kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:36 +0000 UTC Normal Pod agent-as-sidecar-86b878cb8c-p95kv.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:37 +0000 UTC Normal Pod 01-find-service-56mk4 Binding Scheduled Successfully assigned kuttl-test-workable-kodiak/01-find-service-56mk4 to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:37 +0000 UTC Normal Pod 01-find-service-56mk4 AddedInterface Add eth0 [10.131.0.59/23] from ovn-kubernetes multus logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:37 +0000 UTC Normal Pod 01-find-service-56mk4.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:37 +0000 UTC Normal Pod 01-find-service-56mk4.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:37 +0000 UTC Normal Pod 01-find-service-56mk4.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:37 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-56mk4 job-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:39 +0000 UTC Normal Pod vertx-create-span-sidecar-6cc549fcbd-774tm Binding Scheduled Successfully assigned kuttl-test-workable-kodiak/vertx-create-span-sidecar-6cc549fcbd-774tm to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:39 +0000 UTC Normal Pod vertx-create-span-sidecar-6cc549fcbd-774tm AddedInterface Add eth0 [10.129.2.69/23] from ovn-kubernetes multus logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:39 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6cc549fcbd SuccessfulCreate Created pod: vertx-create-span-sidecar-6cc549fcbd-774tm replicaset-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:39 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-jpj6f replicaset-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:39 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:39 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6cc549fcbd to 1 from 0 deployment-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:40 +0000 UTC Normal Pod vertx-create-span-sidecar-6cc549fcbd-774tm.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:40 +0000 UTC Normal Pod vertx-create-span-sidecar-6cc549fcbd-774tm.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:40 +0000 UTC Normal Pod vertx-create-span-sidecar-6cc549fcbd-774tm.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:40 +0000 UTC Normal Pod vertx-create-span-sidecar-6cc549fcbd-774tm.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:40 +0000 UTC Normal Pod vertx-create-span-sidecar-6cc549fcbd-774tm.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:40 +0000 UTC Normal Pod vertx-create-span-sidecar-6cc549fcbd-774tm.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:48 +0000 UTC Warning Pod vertx-create-span-sidecar-6cc549fcbd-774tm.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.69:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:48 +0000 UTC Warning Pod vertx-create-span-sidecar-6cc549fcbd-774tm.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.69:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:51 +0000 UTC Normal Pod vertx-create-span-sidecar-6cc549fcbd-774tm.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:51 +0000 UTC Warning Pod vertx-create-span-sidecar-6cc549fcbd-774tm.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.69:8080/": read tcp 10.129.2.2:36616->10.129.2.69:8080: read: connection reset by peer kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:51 +0000 UTC Warning Pod vertx-create-span-sidecar-6cc549fcbd-774tm.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.69:8080/": dial tcp 10.129.2.69:8080: connect: connection refused kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:55 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:56 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-68cffc55cd SuccessfulDelete Deleted pod: vertx-create-span-sidecar-68cffc55cd-dml9d replicaset-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:56 +0000 UTC Normal Pod vertx-create-span-sidecar-bbf6f464c-2zns4 Binding Scheduled Successfully assigned kuttl-test-workable-kodiak/vertx-create-span-sidecar-bbf6f464c-2zns4 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:56 +0000 UTC Normal Pod vertx-create-span-sidecar-bbf6f464c-2zns4 AddedInterface Add eth0 [10.128.2.54/23] from ovn-kubernetes multus logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:56 +0000 UTC Normal Pod vertx-create-span-sidecar-bbf6f464c-2zns4.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:56 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-bbf6f464c SuccessfulCreate Created pod: vertx-create-span-sidecar-bbf6f464c-2zns4 replicaset-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:56 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-68cffc55cd to 0 from 1 deployment-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:56 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-bbf6f464c to 1 from 0 deployment-controller logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:57 +0000 UTC Normal Pod vertx-create-span-sidecar-bbf6f464c-2zns4.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:33:58 | sidecar-deployment | 2024-03-04 07:33:57 +0000 UTC Normal Pod vertx-create-span-sidecar-bbf6f464c-2zns4.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:33:58 | sidecar-deployment | Deleting namespace: kuttl-test-workable-kodiak === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (144.96s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.73s) --- PASS: kuttl/harness/sidecar-namespace (63.47s) --- PASS: kuttl/harness/sidecar-skip-webhook (16.54s) --- PASS: kuttl/harness/sidecar-deployment (59.17s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name sidecar --report --output /logs/artifacts/sidecar.xml ./artifacts/kuttl-report.xml time="2024-03-04T07:34:05Z" level=debug msg="Setting a new name for the test suites" time="2024-03-04T07:34:05Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-03-04T07:34:05Z" level=debug msg="normalizing test case names" time="2024-03-04T07:34:05Z" level=debug msg="sidecar/artifacts -> sidecar_artifacts" time="2024-03-04T07:34:05Z" level=debug msg="sidecar/sidecar-namespace -> sidecar_sidecar_namespace" time="2024-03-04T07:34:05Z" level=debug msg="sidecar/sidecar-skip-webhook -> sidecar_sidecar_skip_webhook" time="2024-03-04T07:34:05Z" level=debug msg="sidecar/sidecar-deployment -> sidecar_sidecar_deployment" +------------------------------+--------+ | NAME | RESULT | +------------------------------+--------+ | sidecar_artifacts | passed | | sidecar_sidecar_namespace | passed | | sidecar_sidecar_skip_webhook | passed | | sidecar_sidecar_deployment | passed | +------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh streaming false true + '[' 3 -ne 3 ']' + test_suite_name=streaming + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/streaming.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-streaming make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ SKIP_ES_EXTERNAL=true \ ./tests/e2e/streaming/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-02-29-062601 True False 46m Cluster version is 4.16.0-0.nightly-2024-02-29-062601' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-02-29-062601 True False 46m Cluster version is 4.16.0-0.nightly-2024-02-29-062601' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/streaming/render.sh ++ export SUITE_DIR=./tests/e2e/streaming ++ SUITE_DIR=./tests/e2e/streaming ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/streaming ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + '[' false = true ']' + start_test streaming-simple + '[' 1 -ne 1 ']' + test_name=streaming-simple + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-simple' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-simple\e[0m' Rendering files for test streaming-simple + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + mkdir -p streaming-simple + cd streaming-simple + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/streaming-jaeger-assert.yaml.template -o ./04-assert.yaml + render_smoke_test simple-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=simple-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + export JAEGER_NAME=simple-streaming + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-tls + '[' 1 -ne 1 ']' + test_name=streaming-with-tls + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-tls' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-tls\e[0m' Rendering files for test streaming-with-tls + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-simple + '[' streaming-simple '!=' _build ']' + cd .. + mkdir -p streaming-with-tls + cd streaming-with-tls + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + render_smoke_test tls-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=tls-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + export JAEGER_NAME=tls-streaming + JAEGER_NAME=tls-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-autoprovisioning-autoscale + '[' 1 -ne 1 ']' + test_name=streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-autoprovisioning-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-autoprovisioning-autoscale\e[0m' Rendering files for test streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-tls + '[' streaming-with-tls '!=' _build ']' + cd .. + mkdir -p streaming-with-autoprovisioning-autoscale + cd streaming-with-autoprovisioning-autoscale + '[' true = true ']' + rm ./00-install.yaml ./00-assert.yaml + render_install_elasticsearch upstream 01 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=01 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./01-assert.yaml + jaeger_name=auto-provisioned + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="20Mi"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="500m"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.autoscale=true ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.minReplicas=1 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.maxReplicas=2 ./02-install.yaml + render_assert_kafka true auto-provisioned 03 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provisioned + test_step=03 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./03-assert.yaml ++ expr 03 + 1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./04-assert.yaml ++ expr 03 + 2 + CLUSTER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./05-assert.yaml + version_lt 1.27 1.23 ++ echo 1.27 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.27 '!=' 1.27 + rm ./08-assert.yaml + skip_test streaming-with-tls 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-with-tls + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-autoprovisioning-autoscale + '[' streaming-with-autoprovisioning-autoscale '!=' _build ']' + cd .. + rm -rf streaming-with-tls + warning 'streaming-with-tls: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-with-tls: This test is flaky in Prow CI\e[0m' WAR: streaming-with-tls: This test is flaky in Prow CI + skip_test streaming-simple 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-simple + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + rm -rf streaming-simple + warning 'streaming-simple: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-simple: This test is flaky in Prow CI\e[0m' WAR: streaming-simple: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running streaming E2E tests' Running streaming E2E tests + cd tests/e2e/streaming/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2507913876 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 2 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/streaming-with-autoprovisioning-autoscale === PAUSE kuttl/harness/streaming-with-autoprovisioning-autoscale === CONT kuttl/harness/artifacts logger.go:42: 07:34:17 | artifacts | Creating namespace: kuttl-test-summary-shrew logger.go:42: 07:34:17 | artifacts | artifacts events from ns kuttl-test-summary-shrew: logger.go:42: 07:34:17 | artifacts | Deleting namespace: kuttl-test-summary-shrew === CONT kuttl/harness/streaming-with-autoprovisioning-autoscale logger.go:42: 07:34:23 | streaming-with-autoprovisioning-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:34:23 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:34:23 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:34:23 | streaming-with-autoprovisioning-autoscale | Creating namespace: kuttl-test-wondrous-deer logger.go:42: 07:34:23 | streaming-with-autoprovisioning-autoscale/1-install | starting test step 1-install logger.go:42: 07:34:23 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc create sa deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:34:23 | streaming-with-autoprovisioning-autoscale/1-install | serviceaccount/deploy-elasticsearch created logger.go:42: 07:34:23 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc adm policy add-scc-to-user privileged -z deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:34:23 | streaming-with-autoprovisioning-autoscale/1-install | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:privileged added: "deploy-elasticsearch" logger.go:42: 07:34:23 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 6] logger.go:42: 07:34:29 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_0.yml -n $NAMESPACE] logger.go:42: 07:34:29 | streaming-with-autoprovisioning-autoscale/1-install | statefulset.apps/elasticsearch created logger.go:42: 07:34:29 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 3] logger.go:42: 07:34:32 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_1.yml -n $NAMESPACE] logger.go:42: 07:34:33 | streaming-with-autoprovisioning-autoscale/1-install | service/elasticsearch created logger.go:42: 07:34:50 | streaming-with-autoprovisioning-autoscale/1-install | test step completed 1-install logger.go:42: 07:34:50 | streaming-with-autoprovisioning-autoscale/2-install | starting test step 2-install logger.go:42: 07:34:50 | streaming-with-autoprovisioning-autoscale/2-install | Jaeger:kuttl-test-wondrous-deer/auto-provisioned created logger.go:42: 07:34:50 | streaming-with-autoprovisioning-autoscale/2-install | test step completed 2-install logger.go:42: 07:34:50 | streaming-with-autoprovisioning-autoscale/3- | starting test step 3- logger.go:42: 07:35:34 | streaming-with-autoprovisioning-autoscale/3- | test step completed 3- logger.go:42: 07:35:34 | streaming-with-autoprovisioning-autoscale/4- | starting test step 4- logger.go:42: 07:36:15 | streaming-with-autoprovisioning-autoscale/4- | test step completed 4- logger.go:42: 07:36:15 | streaming-with-autoprovisioning-autoscale/5- | starting test step 5- logger.go:42: 07:36:38 | streaming-with-autoprovisioning-autoscale/5- | test step completed 5- logger.go:42: 07:36:38 | streaming-with-autoprovisioning-autoscale/6- | starting test step 6- logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale/6- | test step completed 6- logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale/7- | starting test step 7- logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale/7- | test step completed 7- logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | streaming-with-autoprovisioning-autoscale events from ns kuttl-test-wondrous-deer: logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:34:30 +0000 UTC Normal Pod elasticsearch-0 Binding Scheduled Successfully assigned kuttl-test-wondrous-deer/elasticsearch-0 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:34:30 +0000 UTC Normal Pod elasticsearch-0 AddedInterface Add eth0 [10.128.2.55/23] from ovn-kubernetes multus logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:34:30 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulling Pulling image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:34:30 +0000 UTC Normal StatefulSet.apps elasticsearch SuccessfulCreate create Pod elasticsearch-0 in StatefulSet elasticsearch successful statefulset-controller logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:34:38 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulled Successfully pulled image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" in 8.043s (8.043s including waiting) kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:34:38 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:34:38 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:34:45 +0000 UTC Warning Pod elasticsearch-0.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Get "http://10.128.2.55:9200/": dial tcp 10.128.2.55:9200: connect: connection refused kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:34:54 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:34:55 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-zookeeper NoPods No matching pods found controllermanager logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:34:55 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-wondrous-deer/data-auto-provisioned-zookeeper-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-6fb5bccdfd-qfcn5_da46f58a-d33b-4625-90a0-b0bb88f9cb85 logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:34:55 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:34:59 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ProvisioningSucceeded Successfully provisioned volume pvc-bfd456ef-cc09-4fb3-b83d-fd47fbc0ba85 ebs.csi.aws.com_aws-ebs-csi-driver-controller-6fb5bccdfd-qfcn5_da46f58a-d33b-4625-90a0-b0bb88f9cb85 logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:00 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 Binding Scheduled Successfully assigned kuttl-test-wondrous-deer/auto-provisioned-zookeeper-0 to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:02 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-bfd456ef-cc09-4fb3-b83d-fd47fbc0ba85" attachdetach-controller logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:04 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 AddedInterface Add eth0 [10.129.2.70/23] from ovn-kubernetes multus logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:04 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:99d804bbfd32d006c58dec2e4b78b501834d62f5958c68774b233dd412163456" kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:11 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:99d804bbfd32d006c58dec2e4b78b501834d62f5958c68774b233dd412163456" in 6.626s (6.626s including waiting) kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:11 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Created Created container zookeeper kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:11 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Started Started container zookeeper kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:36 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-kafka NoPods No matching pods found controllermanager logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:36 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:36 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-wondrous-deer/data-0-auto-provisioned-kafka-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-6fb5bccdfd-qfcn5_da46f58a-d33b-4625-90a0-b0bb88f9cb85 logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:36 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:40 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ProvisioningSucceeded Successfully provisioned volume pvc-c5b7efd8-1620-4e4d-b892-2a6f1e9a05cc ebs.csi.aws.com_aws-ebs-csi-driver-controller-6fb5bccdfd-qfcn5_da46f58a-d33b-4625-90a0-b0bb88f9cb85 logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:41 +0000 UTC Normal Pod auto-provisioned-kafka-0 Binding Scheduled Successfully assigned kuttl-test-wondrous-deer/auto-provisioned-kafka-0 to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:43 +0000 UTC Normal Pod auto-provisioned-kafka-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-c5b7efd8-1620-4e4d-b892-2a6f1e9a05cc" attachdetach-controller logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:44 +0000 UTC Normal Pod auto-provisioned-kafka-0 AddedInterface Add eth0 [10.131.0.60/23] from ovn-kubernetes multus logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:44 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:99d804bbfd32d006c58dec2e4b78b501834d62f5958c68774b233dd412163456" kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:51 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:99d804bbfd32d006c58dec2e4b78b501834d62f5958c68774b233dd412163456" in 6.982s (6.982s including waiting) kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:51 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Created Created container kafka kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:35:51 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Started Started container kafka kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:16 +0000 UTC Normal Pod auto-provisioned-entity-operator-7d7767c59f-wjq99 Binding Scheduled Successfully assigned kuttl-test-wondrous-deer/auto-provisioned-entity-operator-7d7767c59f-wjq99 to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:16 +0000 UTC Normal ReplicaSet.apps auto-provisioned-entity-operator-7d7767c59f SuccessfulCreate Created pod: auto-provisioned-entity-operator-7d7767c59f-wjq99 replicaset-controller logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:16 +0000 UTC Normal Deployment.apps auto-provisioned-entity-operator ScalingReplicaSet Scaled up replica set auto-provisioned-entity-operator-7d7767c59f to 1 deployment-controller logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:17 +0000 UTC Normal Pod auto-provisioned-entity-operator-7d7767c59f-wjq99 AddedInterface Add eth0 [10.129.2.71/23] from ovn-kubernetes multus logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:17 +0000 UTC Normal Pod auto-provisioned-entity-operator-7d7767c59f-wjq99.spec.containers{topic-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:5059b9ac793f91de2606ef724a2bb7d6518e75a053cc662665459eafa5f4d074" already present on machine kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:17 +0000 UTC Normal Pod auto-provisioned-entity-operator-7d7767c59f-wjq99.spec.containers{topic-operator} Created Created container topic-operator kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:17 +0000 UTC Normal Pod auto-provisioned-entity-operator-7d7767c59f-wjq99.spec.containers{topic-operator} Started Started container topic-operator kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:17 +0000 UTC Normal Pod auto-provisioned-entity-operator-7d7767c59f-wjq99.spec.containers{user-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:5059b9ac793f91de2606ef724a2bb7d6518e75a053cc662665459eafa5f4d074" already present on machine kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:17 +0000 UTC Normal Pod auto-provisioned-entity-operator-7d7767c59f-wjq99.spec.containers{user-operator} Created Created container user-operator kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:17 +0000 UTC Normal Pod auto-provisioned-entity-operator-7d7767c59f-wjq99.spec.containers{user-operator} Started Started container user-operator kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:17 +0000 UTC Normal Pod auto-provisioned-entity-operator-7d7767c59f-wjq99.spec.containers{tls-sidecar} Pulled Container image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:99d804bbfd32d006c58dec2e4b78b501834d62f5958c68774b233dd412163456" already present on machine kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:17 +0000 UTC Normal Pod auto-provisioned-entity-operator-7d7767c59f-wjq99.spec.containers{tls-sidecar} Created Created container tls-sidecar kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:17 +0000 UTC Normal Pod auto-provisioned-entity-operator-7d7767c59f-wjq99.spec.containers{tls-sidecar} Started Started container tls-sidecar kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:39 +0000 UTC Normal Pod auto-provisioned-collector-5697699c5d-tprx4 Binding Scheduled Successfully assigned kuttl-test-wondrous-deer/auto-provisioned-collector-5697699c5d-tprx4 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:39 +0000 UTC Normal ReplicaSet.apps auto-provisioned-collector-5697699c5d SuccessfulCreate Created pod: auto-provisioned-collector-5697699c5d-tprx4 replicaset-controller logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:39 +0000 UTC Normal Deployment.apps auto-provisioned-collector ScalingReplicaSet Scaled up replica set auto-provisioned-collector-5697699c5d to 1 deployment-controller logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:39 +0000 UTC Normal ReplicaSet.apps auto-provisioned-ingester-66bc96545d SuccessfulCreate Created pod: auto-provisioned-ingester-66bc96545d-k479l replicaset-controller logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:39 +0000 UTC Normal Deployment.apps auto-provisioned-ingester ScalingReplicaSet Scaled up replica set auto-provisioned-ingester-66bc96545d to 1 deployment-controller logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:39 +0000 UTC Normal Pod auto-provisioned-query-7bcf7d68ff-tz68x Binding Scheduled Successfully assigned kuttl-test-wondrous-deer/auto-provisioned-query-7bcf7d68ff-tz68x to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:39 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-7bcf7d68ff SuccessfulCreate Created pod: auto-provisioned-query-7bcf7d68ff-tz68x replicaset-controller logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:39 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled up replica set auto-provisioned-query-7bcf7d68ff to 1 deployment-controller logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:40 +0000 UTC Warning Pod auto-provisioned-collector-5697699c5d-tprx4 FailedMount MountVolume.SetUp failed for volume "auto-provisioned-collector-tls-config-volume" : secret "auto-provisioned-collector-headless-tls" not found kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:40 +0000 UTC Normal Pod auto-provisioned-ingester-66bc96545d-k479l Binding Scheduled Successfully assigned kuttl-test-wondrous-deer/auto-provisioned-ingester-66bc96545d-k479l to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:40 +0000 UTC Normal Pod auto-provisioned-ingester-66bc96545d-k479l AddedInterface Add eth0 [10.128.2.57/23] from ovn-kubernetes multus logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:40 +0000 UTC Normal Pod auto-provisioned-ingester-66bc96545d-k479l.spec.containers{jaeger-ingester} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:158465cf0c6d620a1271482e2f9821d3af0c0548f5b317641034ee711748bc9c" kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:40 +0000 UTC Normal Pod auto-provisioned-query-7bcf7d68ff-tz68x AddedInterface Add eth0 [10.131.0.61/23] from ovn-kubernetes multus logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:40 +0000 UTC Normal Pod auto-provisioned-query-7bcf7d68ff-tz68x.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:40 +0000 UTC Normal Pod auto-provisioned-query-7bcf7d68ff-tz68x.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:40 +0000 UTC Normal Pod auto-provisioned-query-7bcf7d68ff-tz68x.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:40 +0000 UTC Normal Pod auto-provisioned-query-7bcf7d68ff-tz68x.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:40 +0000 UTC Normal Pod auto-provisioned-query-7bcf7d68ff-tz68x.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:40 +0000 UTC Normal Pod auto-provisioned-query-7bcf7d68ff-tz68x.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:40 +0000 UTC Normal Pod auto-provisioned-query-7bcf7d68ff-tz68x.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:40 +0000 UTC Normal Pod auto-provisioned-query-7bcf7d68ff-tz68x.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:40 +0000 UTC Normal Pod auto-provisioned-query-7bcf7d68ff-tz68x.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:41 +0000 UTC Normal Pod auto-provisioned-collector-5697699c5d-tprx4 AddedInterface Add eth0 [10.128.2.56/23] from ovn-kubernetes multus logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:41 +0000 UTC Normal Pod auto-provisioned-collector-5697699c5d-tprx4.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:44 +0000 UTC Normal Pod auto-provisioned-collector-5697699c5d-tprx4.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" in 3.713s (3.713s including waiting) kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:44 +0000 UTC Normal Pod auto-provisioned-collector-5697699c5d-tprx4.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:45 +0000 UTC Normal Pod auto-provisioned-collector-5697699c5d-tprx4.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:45 +0000 UTC Normal Pod auto-provisioned-ingester-66bc96545d-k479l.spec.containers{jaeger-ingester} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:158465cf0c6d620a1271482e2f9821d3af0c0548f5b317641034ee711748bc9c" in 5.347s (5.347s including waiting) kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:46 +0000 UTC Normal Pod auto-provisioned-ingester-66bc96545d-k479l.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | 2024-03-04 07:36:46 +0000 UTC Normal Pod auto-provisioned-ingester-66bc96545d-k479l.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 07:36:48 | streaming-with-autoprovisioning-autoscale | Deleting namespace: kuttl-test-wondrous-deer === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (191.02s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.75s) --- PASS: kuttl/harness/streaming-with-autoprovisioning-autoscale (185.23s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name streaming --report --output /logs/artifacts/streaming.xml ./artifacts/kuttl-report.xml time="2024-03-04T07:37:28Z" level=debug msg="Setting a new name for the test suites" time="2024-03-04T07:37:28Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-03-04T07:37:28Z" level=debug msg="normalizing test case names" time="2024-03-04T07:37:28Z" level=debug msg="streaming/artifacts -> streaming_artifacts" time="2024-03-04T07:37:28Z" level=debug msg="streaming/streaming-with-autoprovisioning-autoscale -> streaming_streaming_with_autoprovisioning_autoscale" +-----------------------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------------------+--------+ | streaming_artifacts | passed | | streaming_streaming_with_autoprovisioning_autoscale | passed | +-----------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh ui false true + '[' 3 -ne 3 ']' + test_suite_name=ui + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/ui.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-ui make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true ./tests/e2e/ui/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-02-29-062601 True False 49m Cluster version is 4.16.0-0.nightly-2024-02-29-062601' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-02-29-062601 True False 49m Cluster version is 4.16.0-0.nightly-2024-02-29-062601' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/ui/render.sh ++ export SUITE_DIR=./tests/e2e/ui ++ SUITE_DIR=./tests/e2e/ui ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/ui ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test allinone + '[' 1 -ne 1 ']' + test_name=allinone + echo =========================================================================== =========================================================================== + info 'Rendering files for test allinone' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test allinone\e[0m' Rendering files for test allinone + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build + '[' _build '!=' _build ']' + mkdir -p allinone + cd allinone + export GET_URL_COMMAND + export URL + export JAEGER_NAME=all-in-one-ui + JAEGER_NAME=all-in-one-ui + '[' true = true ']' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./01-curl.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./04-test-ui-config.yaml + start_test production + '[' 1 -ne 1 ']' + test_name=production + echo =========================================================================== =========================================================================== + info 'Rendering files for test production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test production\e[0m' Rendering files for test production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build/allinone + '[' allinone '!=' _build ']' + cd .. + mkdir -p production + cd production + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + [[ true = true ]] + [[ true = true ]] + render_install_jaeger production-ui production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + '[' true = true ']' + INSECURE=true + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-forbbiden-access.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-curl.yaml + INSECURE=true + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./05-check-disabled-security.yaml + ASSERT_PRESENT=false + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./06-check-NO-gaID.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./08-check-gaID.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running ui E2E tests' Running ui E2E tests + cd tests/e2e/ui/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2507913876 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 3 tests === RUN kuttl/harness === RUN kuttl/harness/allinone === PAUSE kuttl/harness/allinone === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/production === PAUSE kuttl/harness/production === CONT kuttl/harness/allinone logger.go:42: 07:37:35 | allinone | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:37:35 | allinone | Creating namespace: kuttl-test-legal-gator logger.go:42: 07:37:35 | allinone/0-install | starting test step 0-install logger.go:42: 07:37:35 | allinone/0-install | Jaeger:kuttl-test-legal-gator/all-in-one-ui created logger.go:42: 07:37:39 | allinone/0-install | test step completed 0-install logger.go:42: 07:37:39 | allinone/1-curl | starting test step 1-curl logger.go:42: 07:37:39 | allinone/1-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 07:37:39 | allinone/1-curl | Checking the Ingress host value was populated logger.go:42: 07:37:39 | allinone/1-curl | Try number 0 logger.go:42: 07:37:39 | allinone/1-curl | Hostname is all-in-one-ui-kuttl-test-legal-gator.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com logger.go:42: 07:37:39 | allinone/1-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE all-in-one-ui] logger.go:42: 07:37:39 | allinone/1-curl | Checking an expected HTTP response logger.go:42: 07:37:39 | allinone/1-curl | Running in OpenShift logger.go:42: 07:37:39 | allinone/1-curl | User not provided. Getting the token... logger.go:42: 07:37:41 | allinone/1-curl | Warning: resource jaegers/all-in-one-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:37:47 | allinone/1-curl | Try number 1/30 the https://all-in-one-ui-kuttl-test-legal-gator.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:37:47 | allinone/1-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:37:47 | allinone/1-curl | Try number 2/30 the https://all-in-one-ui-kuttl-test-legal-gator.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:37:47 | allinone/1-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 07:37:57 | allinone/1-curl | Try number 3/30 the https://all-in-one-ui-kuttl-test-legal-gator.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:37:57 | allinone/1-curl | curl response asserted properly logger.go:42: 07:37:57 | allinone/1-curl | test step completed 1-curl logger.go:42: 07:37:57 | allinone/2-delete | starting test step 2-delete logger.go:42: 07:37:57 | allinone/2-delete | Jaeger:kuttl-test-legal-gator/all-in-one-ui created logger.go:42: 07:37:57 | allinone/2-delete | test step completed 2-delete logger.go:42: 07:37:57 | allinone/3-install | starting test step 3-install logger.go:42: 07:37:57 | allinone/3-install | Jaeger:kuttl-test-legal-gator/all-in-one-ui updated logger.go:42: 07:37:57 | allinone/3-install | test step completed 3-install logger.go:42: 07:37:57 | allinone/4-test-ui-config | starting test step 4-test-ui-config logger.go:42: 07:37:57 | allinone/4-test-ui-config | running command: [./ensure-ingress-host.sh] logger.go:42: 07:37:57 | allinone/4-test-ui-config | Checking the Ingress host value was populated logger.go:42: 07:37:57 | allinone/4-test-ui-config | Try number 0 logger.go:42: 07:37:57 | allinone/4-test-ui-config | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 07:37:57 | allinone/4-test-ui-config | template was: logger.go:42: 07:37:57 | allinone/4-test-ui-config | {.items[0].status.ingress[0].host} logger.go:42: 07:37:57 | allinone/4-test-ui-config | object given to jsonpath engine was: logger.go:42: 07:37:57 | allinone/4-test-ui-config | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 07:37:57 | allinone/4-test-ui-config | logger.go:42: 07:37:57 | allinone/4-test-ui-config | logger.go:42: 07:38:07 | allinone/4-test-ui-config | Try number 1 logger.go:42: 07:38:08 | allinone/4-test-ui-config | Hostname is all-in-one-ui-kuttl-test-legal-gator.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com logger.go:42: 07:38:08 | allinone/4-test-ui-config | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:38:08 | allinone/4-test-ui-config | time="2024-03-04T07:38:08Z" level=info msg="Querying https://all-in-one-ui-kuttl-test-legal-gator.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:38:08 | allinone/4-test-ui-config | time="2024-03-04T07:38:08Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:38:08 | allinone/4-test-ui-config | time="2024-03-04T07:38:08Z" level=info msg="Polling to https://all-in-one-ui-kuttl-test-legal-gator.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:38:08 | allinone/4-test-ui-config | time="2024-03-04T07:38:08Z" level=info msg="Doing request number 0" logger.go:42: 07:38:08 | allinone/4-test-ui-config | time="2024-03-04T07:38:08Z" level=info msg="Content found and asserted!" logger.go:42: 07:38:08 | allinone/4-test-ui-config | time="2024-03-04T07:38:08Z" level=info msg="Success!" logger.go:42: 07:38:08 | allinone/4-test-ui-config | test step completed 4-test-ui-config logger.go:42: 07:38:08 | allinone | allinone events from ns kuttl-test-legal-gator: logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:39 +0000 UTC Normal Pod all-in-one-ui-767595cfd7-4zftv Binding Scheduled Successfully assigned kuttl-test-legal-gator/all-in-one-ui-767595cfd7-4zftv to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:39 +0000 UTC Normal Pod all-in-one-ui-767595cfd7-4zftv AddedInterface Add eth0 [10.128.2.58/23] from ovn-kubernetes multus logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:39 +0000 UTC Normal Pod all-in-one-ui-767595cfd7-4zftv.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:39 +0000 UTC Normal Pod all-in-one-ui-767595cfd7-4zftv.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:39 +0000 UTC Normal Pod all-in-one-ui-767595cfd7-4zftv.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:39 +0000 UTC Normal Pod all-in-one-ui-767595cfd7-4zftv.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:39 +0000 UTC Normal Pod all-in-one-ui-767595cfd7-4zftv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:39 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-767595cfd7 SuccessfulCreate Created pod: all-in-one-ui-767595cfd7-4zftv replicaset-controller logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:39 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-767595cfd7 to 1 deployment-controller logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:40 +0000 UTC Normal Pod all-in-one-ui-767595cfd7-4zftv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:43 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-767595cfd7 SuccessfulDelete Deleted pod: all-in-one-ui-767595cfd7-4zftv replicaset-controller logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:43 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-767595cfd7 to 0 from 1 deployment-controller logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:44 +0000 UTC Normal Pod all-in-one-ui-767595cfd7-4zftv.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:44 +0000 UTC Normal Pod all-in-one-ui-767595cfd7-4zftv.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:45 +0000 UTC Normal Pod all-in-one-ui-c48cd4b55-l7cf6 Binding Scheduled Successfully assigned kuttl-test-legal-gator/all-in-one-ui-c48cd4b55-l7cf6 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:45 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-c48cd4b55 SuccessfulCreate Created pod: all-in-one-ui-c48cd4b55-l7cf6 replicaset-controller logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:45 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-c48cd4b55 to 1 deployment-controller logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:46 +0000 UTC Normal Pod all-in-one-ui-c48cd4b55-l7cf6 AddedInterface Add eth0 [10.128.2.59/23] from ovn-kubernetes multus logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:46 +0000 UTC Normal Pod all-in-one-ui-c48cd4b55-l7cf6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:46 +0000 UTC Normal Pod all-in-one-ui-c48cd4b55-l7cf6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:46 +0000 UTC Normal Pod all-in-one-ui-c48cd4b55-l7cf6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:46 +0000 UTC Normal Pod all-in-one-ui-c48cd4b55-l7cf6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:46 +0000 UTC Normal Pod all-in-one-ui-c48cd4b55-l7cf6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:46 +0000 UTC Normal Pod all-in-one-ui-c48cd4b55-l7cf6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:57 +0000 UTC Normal Pod all-in-one-ui-c48cd4b55-l7cf6.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:38:08 | allinone | 2024-03-04 07:37:57 +0000 UTC Normal Pod all-in-one-ui-c48cd4b55-l7cf6.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:38:08 | allinone | 2024-03-04 07:38:03 +0000 UTC Normal Pod all-in-one-ui-6fb74c6b4b-6wp4d Binding Scheduled Successfully assigned kuttl-test-legal-gator/all-in-one-ui-6fb74c6b4b-6wp4d to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:38:08 | allinone | 2024-03-04 07:38:03 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-6fb74c6b4b SuccessfulCreate Created pod: all-in-one-ui-6fb74c6b4b-6wp4d replicaset-controller logger.go:42: 07:38:08 | allinone | 2024-03-04 07:38:03 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-6fb74c6b4b to 1 deployment-controller logger.go:42: 07:38:08 | allinone | 2024-03-04 07:38:04 +0000 UTC Normal Pod all-in-one-ui-6fb74c6b4b-6wp4d AddedInterface Add eth0 [10.128.2.60/23] from ovn-kubernetes multus logger.go:42: 07:38:08 | allinone | 2024-03-04 07:38:04 +0000 UTC Normal Pod all-in-one-ui-6fb74c6b4b-6wp4d.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:38:08 | allinone | 2024-03-04 07:38:04 +0000 UTC Normal Pod all-in-one-ui-6fb74c6b4b-6wp4d.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:38:08 | allinone | 2024-03-04 07:38:04 +0000 UTC Normal Pod all-in-one-ui-6fb74c6b4b-6wp4d.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:38:08 | allinone | Deleting namespace: kuttl-test-legal-gator === CONT kuttl/harness/production logger.go:42: 07:38:14 | production | Ignoring add-tracking-id.yaml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:38:14 | production | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:38:14 | production | Creating namespace: kuttl-test-calm-hen logger.go:42: 07:38:14 | production/1-install | starting test step 1-install logger.go:42: 07:38:14 | production/1-install | Jaeger:kuttl-test-calm-hen/production-ui created logger.go:42: 07:38:50 | production/1-install | test step completed 1-install logger.go:42: 07:38:50 | production/2-check-forbbiden-access | starting test step 2-check-forbbiden-access logger.go:42: 07:38:50 | production/2-check-forbbiden-access | running command: [./ensure-ingress-host.sh] logger.go:42: 07:38:50 | production/2-check-forbbiden-access | Checking the Ingress host value was populated logger.go:42: 07:38:50 | production/2-check-forbbiden-access | Try number 0 logger.go:42: 07:38:50 | production/2-check-forbbiden-access | Hostname is production-ui-kuttl-test-calm-hen.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com logger.go:42: 07:38:50 | production/2-check-forbbiden-access | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE production-ui] logger.go:42: 07:38:50 | production/2-check-forbbiden-access | Checking an expected HTTP response logger.go:42: 07:38:50 | production/2-check-forbbiden-access | Running in OpenShift logger.go:42: 07:38:50 | production/2-check-forbbiden-access | Not using any secret logger.go:42: 07:38:50 | production/2-check-forbbiden-access | Try number 1/30 the https://production-ui-kuttl-test-calm-hen.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:38:50 | production/2-check-forbbiden-access | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:38:50 | production/2-check-forbbiden-access | Try number 2/30 the https://production-ui-kuttl-test-calm-hen.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:38:50 | production/2-check-forbbiden-access | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 07:39:00 | production/2-check-forbbiden-access | Try number 3/30 the https://production-ui-kuttl-test-calm-hen.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:39:00 | production/2-check-forbbiden-access | curl response asserted properly logger.go:42: 07:39:00 | production/2-check-forbbiden-access | test step completed 2-check-forbbiden-access logger.go:42: 07:39:00 | production/3-curl | starting test step 3-curl logger.go:42: 07:39:00 | production/3-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 07:39:00 | production/3-curl | Checking the Ingress host value was populated logger.go:42: 07:39:00 | production/3-curl | Try number 0 logger.go:42: 07:39:00 | production/3-curl | Hostname is production-ui-kuttl-test-calm-hen.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com logger.go:42: 07:39:00 | production/3-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 07:39:00 | production/3-curl | Checking an expected HTTP response logger.go:42: 07:39:00 | production/3-curl | Running in OpenShift logger.go:42: 07:39:00 | production/3-curl | User not provided. Getting the token... logger.go:42: 07:39:01 | production/3-curl | Warning: resource jaegers/production-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:39:08 | production/3-curl | Try number 1/30 the https://production-ui-kuttl-test-calm-hen.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:39:08 | production/3-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:39:08 | production/3-curl | Try number 2/30 the https://production-ui-kuttl-test-calm-hen.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:39:08 | production/3-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 07:39:18 | production/3-curl | Try number 3/30 the https://production-ui-kuttl-test-calm-hen.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:39:18 | production/3-curl | curl response asserted properly logger.go:42: 07:39:18 | production/3-curl | test step completed 3-curl logger.go:42: 07:39:18 | production/4-install | starting test step 4-install logger.go:42: 07:39:18 | production/4-install | Jaeger:kuttl-test-calm-hen/production-ui updated logger.go:42: 07:39:18 | production/4-install | test step completed 4-install logger.go:42: 07:39:18 | production/5-check-disabled-security | starting test step 5-check-disabled-security logger.go:42: 07:39:18 | production/5-check-disabled-security | running command: [./ensure-ingress-host.sh] logger.go:42: 07:39:18 | production/5-check-disabled-security | Checking the Ingress host value was populated logger.go:42: 07:39:18 | production/5-check-disabled-security | Try number 0 logger.go:42: 07:39:18 | production/5-check-disabled-security | Hostname is production-ui-kuttl-test-calm-hen.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com logger.go:42: 07:39:18 | production/5-check-disabled-security | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 07:39:18 | production/5-check-disabled-security | Checking an expected HTTP response logger.go:42: 07:39:18 | production/5-check-disabled-security | Running in OpenShift logger.go:42: 07:39:18 | production/5-check-disabled-security | Not using any secret logger.go:42: 07:39:18 | production/5-check-disabled-security | Try number 1/30 the https://production-ui-kuttl-test-calm-hen.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:39:18 | production/5-check-disabled-security | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:39:18 | production/5-check-disabled-security | Try number 2/30 the https://production-ui-kuttl-test-calm-hen.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:39:18 | production/5-check-disabled-security | HTTP response is 403. 200 expected. Waiting 10 s logger.go:42: 07:39:28 | production/5-check-disabled-security | Try number 3/30 the https://production-ui-kuttl-test-calm-hen.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search logger.go:42: 07:39:28 | production/5-check-disabled-security | curl response asserted properly logger.go:42: 07:39:28 | production/5-check-disabled-security | test step completed 5-check-disabled-security logger.go:42: 07:39:28 | production/6-check-NO-gaID | starting test step 6-check-NO-gaID logger.go:42: 07:39:28 | production/6-check-NO-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 07:39:28 | production/6-check-NO-gaID | Checking the Ingress host value was populated logger.go:42: 07:39:28 | production/6-check-NO-gaID | Try number 0 logger.go:42: 07:39:28 | production/6-check-NO-gaID | Hostname is production-ui-kuttl-test-calm-hen.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com logger.go:42: 07:39:28 | production/6-check-NO-gaID | running command: [sh -c ASSERT_PRESENT=false EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:39:28 | production/6-check-NO-gaID | time="2024-03-04T07:39:28Z" level=info msg="Querying https://production-ui-kuttl-test-calm-hen.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:39:28 | production/6-check-NO-gaID | time="2024-03-04T07:39:28Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:39:28 | production/6-check-NO-gaID | time="2024-03-04T07:39:28Z" level=info msg="Polling to https://production-ui-kuttl-test-calm-hen.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:39:28 | production/6-check-NO-gaID | time="2024-03-04T07:39:28Z" level=info msg="Doing request number 0" logger.go:42: 07:39:28 | production/6-check-NO-gaID | time="2024-03-04T07:39:28Z" level=info msg="Content not found and asserted it was not found!" logger.go:42: 07:39:28 | production/6-check-NO-gaID | time="2024-03-04T07:39:28Z" level=info msg="Success!" logger.go:42: 07:39:28 | production/6-check-NO-gaID | test step completed 6-check-NO-gaID logger.go:42: 07:39:28 | production/7-add-tracking-id | starting test step 7-add-tracking-id logger.go:42: 07:39:28 | production/7-add-tracking-id | running command: [sh -c kubectl apply -f add-tracking-id.yaml -n $NAMESPACE] logger.go:42: 07:39:29 | production/7-add-tracking-id | jaeger.jaegertracing.io/production-ui configured logger.go:42: 07:39:29 | production/7-add-tracking-id | test step completed 7-add-tracking-id logger.go:42: 07:39:29 | production/8-check-gaID | starting test step 8-check-gaID logger.go:42: 07:39:29 | production/8-check-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 07:39:29 | production/8-check-gaID | Checking the Ingress host value was populated logger.go:42: 07:39:29 | production/8-check-gaID | Try number 0 logger.go:42: 07:39:29 | production/8-check-gaID | Hostname is production-ui-kuttl-test-calm-hen.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com logger.go:42: 07:39:29 | production/8-check-gaID | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:39:29 | production/8-check-gaID | time="2024-03-04T07:39:29Z" level=info msg="Querying https://production-ui-kuttl-test-calm-hen.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:39:29 | production/8-check-gaID | time="2024-03-04T07:39:29Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:39:29 | production/8-check-gaID | time="2024-03-04T07:39:29Z" level=info msg="Polling to https://production-ui-kuttl-test-calm-hen.apps.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:39:29 | production/8-check-gaID | time="2024-03-04T07:39:29Z" level=info msg="Doing request number 0" logger.go:42: 07:39:29 | production/8-check-gaID | time="2024-03-04T07:39:29Z" level=warning msg="Found: false . Assert: true" logger.go:42: 07:39:29 | production/8-check-gaID | time="2024-03-04T07:39:29Z" level=warning msg="The condition of the test function was not accomplished" logger.go:42: 07:39:37 | production/8-check-gaID | time="2024-03-04T07:39:37Z" level=info msg="Doing request number 1" logger.go:42: 07:39:37 | production/8-check-gaID | time="2024-03-04T07:39:37Z" level=info msg="Content found and asserted!" logger.go:42: 07:39:37 | production/8-check-gaID | time="2024-03-04T07:39:37Z" level=info msg="Success!" logger.go:42: 07:39:37 | production/8-check-gaID | test step completed 8-check-gaID logger.go:42: 07:39:37 | production | production events from ns kuttl-test-calm-hen: logger.go:42: 07:39:37 | production | 2024-03-04 07:38:20 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestcalmhenproductionui-1-c6f475686 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestcalmhenproductionui-1-c6f475686f849n replicaset-controller logger.go:42: 07:39:37 | production | 2024-03-04 07:38:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcalmhenproductionui-1-c6f475686f849n Binding Scheduled Successfully assigned kuttl-test-calm-hen/elasticsearch-cdm-kuttltestcalmhenproductionui-1-c6f475686f849n to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:39:37 | production | 2024-03-04 07:38:20 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcalmhenproductionui-1-c6f475686f849n FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:20 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestcalmhenproductionui-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestcalmhenproductionui-1-c6f475686 to 1 deployment-controller logger.go:42: 07:39:37 | production | 2024-03-04 07:38:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcalmhenproductionui-1-c6f475686f849n AddedInterface Add eth0 [10.128.2.61/23] from ovn-kubernetes multus logger.go:42: 07:39:37 | production | 2024-03-04 07:38:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcalmhenproductionui-1-c6f475686f849n.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcalmhenproductionui-1-c6f475686f849n.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcalmhenproductionui-1-c6f475686f849n.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcalmhenproductionui-1-c6f475686f849n.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcalmhenproductionui-1-c6f475686f849n.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcalmhenproductionui-1-c6f475686f849n.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:31 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcalmhenproductionui-1-c6f475686f849n.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:36 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcalmhenproductionui-1-c6f475686f849n.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:47 +0000 UTC Normal Pod production-ui-collector-5b88557dd6-q9tt5 Binding Scheduled Successfully assigned kuttl-test-calm-hen/production-ui-collector-5b88557dd6-q9tt5 to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:39:37 | production | 2024-03-04 07:38:47 +0000 UTC Normal Pod production-ui-collector-5b88557dd6-q9tt5 AddedInterface Add eth0 [10.129.2.72/23] from ovn-kubernetes multus logger.go:42: 07:39:37 | production | 2024-03-04 07:38:47 +0000 UTC Normal Pod production-ui-collector-5b88557dd6-q9tt5.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" already present on machine kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:47 +0000 UTC Normal ReplicaSet.apps production-ui-collector-5b88557dd6 SuccessfulCreate Created pod: production-ui-collector-5b88557dd6-q9tt5 replicaset-controller logger.go:42: 07:39:37 | production | 2024-03-04 07:38:47 +0000 UTC Normal Deployment.apps production-ui-collector ScalingReplicaSet Scaled up replica set production-ui-collector-5b88557dd6 to 1 deployment-controller logger.go:42: 07:39:37 | production | 2024-03-04 07:38:47 +0000 UTC Normal Pod production-ui-query-585cbf9879-vkwvk Binding Scheduled Successfully assigned kuttl-test-calm-hen/production-ui-query-585cbf9879-vkwvk to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:39:37 | production | 2024-03-04 07:38:47 +0000 UTC Normal Pod production-ui-query-585cbf9879-vkwvk AddedInterface Add eth0 [10.131.0.62/23] from ovn-kubernetes multus logger.go:42: 07:39:37 | production | 2024-03-04 07:38:47 +0000 UTC Normal ReplicaSet.apps production-ui-query-585cbf9879 SuccessfulCreate Created pod: production-ui-query-585cbf9879-vkwvk replicaset-controller logger.go:42: 07:39:37 | production | 2024-03-04 07:38:47 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-585cbf9879 to 1 deployment-controller logger.go:42: 07:39:37 | production | 2024-03-04 07:38:48 +0000 UTC Normal Pod production-ui-collector-5b88557dd6-q9tt5.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:48 +0000 UTC Normal Pod production-ui-collector-5b88557dd6-q9tt5.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:48 +0000 UTC Normal Pod production-ui-query-585cbf9879-vkwvk.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:48 +0000 UTC Normal Pod production-ui-query-585cbf9879-vkwvk.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:48 +0000 UTC Normal Pod production-ui-query-585cbf9879-vkwvk.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:48 +0000 UTC Normal Pod production-ui-query-585cbf9879-vkwvk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:48 +0000 UTC Normal Pod production-ui-query-585cbf9879-vkwvk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:48 +0000 UTC Normal Pod production-ui-query-585cbf9879-vkwvk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:48 +0000 UTC Normal Pod production-ui-query-585cbf9879-vkwvk.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:48 +0000 UTC Normal Pod production-ui-query-585cbf9879-vkwvk.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:38:48 +0000 UTC Normal Pod production-ui-query-585cbf9879-vkwvk.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:02 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:39:37 | production | 2024-03-04 07:39:02 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:39:37 | production | 2024-03-04 07:39:02 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:39:37 | production | 2024-03-04 07:39:03 +0000 UTC Normal Pod production-ui-query-585cbf9879-vkwvk.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:03 +0000 UTC Normal Pod production-ui-query-585cbf9879-vkwvk.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:03 +0000 UTC Normal Pod production-ui-query-585cbf9879-vkwvk.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:03 +0000 UTC Normal ReplicaSet.apps production-ui-query-585cbf9879 SuccessfulDelete Deleted pod: production-ui-query-585cbf9879-vkwvk replicaset-controller logger.go:42: 07:39:37 | production | 2024-03-04 07:39:03 +0000 UTC Normal ReplicaSet.apps production-ui-query-6bc7769cf6 SuccessfulCreate Created pod: production-ui-query-6bc7769cf6-gl9nw replicaset-controller logger.go:42: 07:39:37 | production | 2024-03-04 07:39:03 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-585cbf9879 to 0 from 1 deployment-controller logger.go:42: 07:39:37 | production | 2024-03-04 07:39:03 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-6bc7769cf6 to 1 deployment-controller logger.go:42: 07:39:37 | production | 2024-03-04 07:39:04 +0000 UTC Normal Pod production-ui-query-6bc7769cf6-gl9nw Binding Scheduled Successfully assigned kuttl-test-calm-hen/production-ui-query-6bc7769cf6-gl9nw to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:39:37 | production | 2024-03-04 07:39:04 +0000 UTC Normal Pod production-ui-query-6bc7769cf6-gl9nw AddedInterface Add eth0 [10.131.0.63/23] from ovn-kubernetes multus logger.go:42: 07:39:37 | production | 2024-03-04 07:39:04 +0000 UTC Normal Pod production-ui-query-6bc7769cf6-gl9nw.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:04 +0000 UTC Normal Pod production-ui-query-6bc7769cf6-gl9nw.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:04 +0000 UTC Normal Pod production-ui-query-6bc7769cf6-gl9nw.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:04 +0000 UTC Normal Pod production-ui-query-6bc7769cf6-gl9nw.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:04 +0000 UTC Normal Pod production-ui-query-6bc7769cf6-gl9nw.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:04 +0000 UTC Normal Pod production-ui-query-6bc7769cf6-gl9nw.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:04 +0000 UTC Normal Pod production-ui-query-6bc7769cf6-gl9nw.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:05 +0000 UTC Normal Pod production-ui-query-6bc7769cf6-gl9nw.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:05 +0000 UTC Normal Pod production-ui-query-6bc7769cf6-gl9nw.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:19 +0000 UTC Normal Pod production-ui-query-6bc7769cf6-gl9nw.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:19 +0000 UTC Normal Pod production-ui-query-6bc7769cf6-gl9nw.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:19 +0000 UTC Normal Pod production-ui-query-6bc7769cf6-gl9nw.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:19 +0000 UTC Normal ReplicaSet.apps production-ui-query-6bc7769cf6 SuccessfulDelete Deleted pod: production-ui-query-6bc7769cf6-gl9nw replicaset-controller logger.go:42: 07:39:37 | production | 2024-03-04 07:39:19 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-6bc7769cf6 to 0 from 1 deployment-controller logger.go:42: 07:39:37 | production | 2024-03-04 07:39:20 +0000 UTC Normal Pod production-ui-query-85cb7f4d74-pmh95 Binding Scheduled Successfully assigned kuttl-test-calm-hen/production-ui-query-85cb7f4d74-pmh95 to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:39:37 | production | 2024-03-04 07:39:20 +0000 UTC Normal ReplicaSet.apps production-ui-query-85cb7f4d74 SuccessfulCreate Created pod: production-ui-query-85cb7f4d74-pmh95 replicaset-controller logger.go:42: 07:39:37 | production | 2024-03-04 07:39:20 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-85cb7f4d74 to 1 deployment-controller logger.go:42: 07:39:37 | production | 2024-03-04 07:39:21 +0000 UTC Normal Pod production-ui-query-85cb7f4d74-pmh95 AddedInterface Add eth0 [10.131.0.64/23] from ovn-kubernetes multus logger.go:42: 07:39:37 | production | 2024-03-04 07:39:21 +0000 UTC Normal Pod production-ui-query-85cb7f4d74-pmh95.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:21 +0000 UTC Normal Pod production-ui-query-85cb7f4d74-pmh95.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:21 +0000 UTC Normal Pod production-ui-query-85cb7f4d74-pmh95.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:21 +0000 UTC Normal Pod production-ui-query-85cb7f4d74-pmh95.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:21 +0000 UTC Normal Pod production-ui-query-85cb7f4d74-pmh95.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:21 +0000 UTC Normal Pod production-ui-query-85cb7f4d74-pmh95.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:30 +0000 UTC Normal Pod production-ui-query-85cb7f4d74-pmh95.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:30 +0000 UTC Normal Pod production-ui-query-85cb7f4d74-pmh95.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:30 +0000 UTC Normal ReplicaSet.apps production-ui-query-85cb7f4d74 SuccessfulDelete Deleted pod: production-ui-query-85cb7f4d74-pmh95 replicaset-controller logger.go:42: 07:39:37 | production | 2024-03-04 07:39:30 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-85cb7f4d74 to 0 from 1 deployment-controller logger.go:42: 07:39:37 | production | 2024-03-04 07:39:31 +0000 UTC Normal Pod production-ui-query-5fbd7bbf49-kpb26 Binding Scheduled Successfully assigned kuttl-test-calm-hen/production-ui-query-5fbd7bbf49-kpb26 to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:39:37 | production | 2024-03-04 07:39:31 +0000 UTC Normal ReplicaSet.apps production-ui-query-5fbd7bbf49 SuccessfulCreate Created pod: production-ui-query-5fbd7bbf49-kpb26 replicaset-controller logger.go:42: 07:39:37 | production | 2024-03-04 07:39:31 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-5fbd7bbf49 to 1 deployment-controller logger.go:42: 07:39:37 | production | 2024-03-04 07:39:32 +0000 UTC Normal Pod production-ui-query-5fbd7bbf49-kpb26 AddedInterface Add eth0 [10.131.0.65/23] from ovn-kubernetes multus logger.go:42: 07:39:37 | production | 2024-03-04 07:39:32 +0000 UTC Normal Pod production-ui-query-5fbd7bbf49-kpb26.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:32 +0000 UTC Normal Pod production-ui-query-5fbd7bbf49-kpb26.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:32 +0000 UTC Normal Pod production-ui-query-5fbd7bbf49-kpb26.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:32 +0000 UTC Normal Pod production-ui-query-5fbd7bbf49-kpb26.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:32 +0000 UTC Normal Pod production-ui-query-5fbd7bbf49-kpb26.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:39:37 | production | 2024-03-04 07:39:32 +0000 UTC Normal Pod production-ui-query-5fbd7bbf49-kpb26.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:39:37 | production | Deleting namespace: kuttl-test-calm-hen === CONT kuttl/harness/artifacts logger.go:42: 07:39:44 | artifacts | Creating namespace: kuttl-test-main-eagle logger.go:42: 07:39:44 | artifacts | artifacts events from ns kuttl-test-main-eagle: logger.go:42: 07:39:44 | artifacts | Deleting namespace: kuttl-test-main-eagle === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (134.62s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/allinone (39.15s) --- PASS: kuttl/harness/production (89.57s) --- PASS: kuttl/harness/artifacts (5.85s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name ui --report --output /logs/artifacts/ui.xml ./artifacts/kuttl-report.xml time="2024-03-04T07:39:50Z" level=debug msg="Setting a new name for the test suites" time="2024-03-04T07:39:50Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-03-04T07:39:50Z" level=debug msg="normalizing test case names" time="2024-03-04T07:39:50Z" level=debug msg="ui/allinone -> ui_allinone" time="2024-03-04T07:39:50Z" level=debug msg="ui/production -> ui_production" time="2024-03-04T07:39:50Z" level=debug msg="ui/artifacts -> ui_artifacts" +---------------+--------+ | NAME | RESULT | +---------------+--------+ | ui_allinone | passed | | ui_production | passed | | ui_artifacts | passed | +---------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh miscellaneous false true + '[' 3 -ne 3 ']' + test_suite_name=miscellaneous + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/miscellaneous.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-miscellaneous make[2]: Entering directory '/tmp/jaeger-tests' SKIP_ES_EXTERNAL=true ./tests/e2e/miscellaneous/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-02-29-062601 True False 51m Cluster version is 4.16.0-0.nightly-2024-02-29-062601' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-02-29-062601 True False 51m Cluster version is 4.16.0-0.nightly-2024-02-29-062601' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/miscellaneous/render.sh ++ export SUITE_DIR=./tests/e2e/miscellaneous ++ SUITE_DIR=./tests/e2e/miscellaneous ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/miscellaneous ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test cassandra-spark 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=cassandra-spark + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf cassandra-spark + warning 'cassandra-spark: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: cassandra-spark: Test not supported in OpenShift\e[0m' WAR: cassandra-spark: Test not supported in OpenShift + start_test collector-autoscale + '[' 1 -ne 1 ']' + test_name=collector-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-autoscale\e[0m' Rendering files for test collector-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p collector-autoscale + cd collector-autoscale + jaeger_name=simple-prod + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + ELASTICSEARCH_NODECOUNT=1 + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.resources.requests.memory="200m"' 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.autoscale=true 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.minReplicas=1 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.maxReplicas=2 01-install.yaml + version_lt 1.27 1.23 ++ echo 1.27 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.27 '!=' 1.27 + rm ./03-assert.yaml + generate_otlp_e2e_tests http + test_protocol=http + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-http\e[0m' Rendering files for test collector-otlp-allinone-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-autoscale + '[' collector-autoscale '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-http + cd collector-otlp-allinone-http + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger http true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-http\e[0m' Rendering files for test collector-otlp-production-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-http + '[' collector-otlp-allinone-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-http + cd collector-otlp-production-http + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger http true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + generate_otlp_e2e_tests grpc + test_protocol=grpc + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-grpc\e[0m' Rendering files for test collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-http + '[' collector-otlp-production-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-grpc + cd collector-otlp-allinone-grpc + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-grpc\e[0m' Rendering files for test collector-otlp-production-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-grpc + '[' collector-otlp-allinone-grpc '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-grpc + cd collector-otlp-production-grpc + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + '[' true = true ']' + skip_test istio 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=istio + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-grpc + '[' collector-otlp-production-grpc '!=' _build ']' + cd .. + rm -rf istio + warning 'istio: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: istio: Test not supported in OpenShift\e[0m' WAR: istio: Test not supported in OpenShift + '[' true = true ']' + skip_test outside-cluster 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=outside-cluster + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf outside-cluster + warning 'outside-cluster: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: outside-cluster: Test not supported in OpenShift\e[0m' WAR: outside-cluster: Test not supported in OpenShift + start_test set-custom-img + '[' 1 -ne 1 ']' + test_name=set-custom-img + echo =========================================================================== =========================================================================== + info 'Rendering files for test set-custom-img' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test set-custom-img\e[0m' Rendering files for test set-custom-img + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p set-custom-img + cd set-custom-img + jaeger_name=my-jaeger + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.image="test"' ./02-install.yaml + '[' true = true ']' + skip_test non-cluster-wide 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=non-cluster-wide + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/set-custom-img + '[' set-custom-img '!=' _build ']' + cd .. + rm -rf non-cluster-wide + warning 'non-cluster-wide: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: non-cluster-wide: Test not supported in OpenShift\e[0m' WAR: non-cluster-wide: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running miscellaneous E2E tests' Running miscellaneous E2E tests + cd tests/e2e/miscellaneous/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2507913876 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-yjw0rx4s-f2afc.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/collector-autoscale === PAUSE kuttl/harness/collector-autoscale === RUN kuttl/harness/collector-otlp-allinone-grpc === PAUSE kuttl/harness/collector-otlp-allinone-grpc === RUN kuttl/harness/collector-otlp-allinone-http === PAUSE kuttl/harness/collector-otlp-allinone-http === RUN kuttl/harness/collector-otlp-production-grpc === PAUSE kuttl/harness/collector-otlp-production-grpc === RUN kuttl/harness/collector-otlp-production-http === PAUSE kuttl/harness/collector-otlp-production-http === RUN kuttl/harness/set-custom-img === PAUSE kuttl/harness/set-custom-img === CONT kuttl/harness/artifacts logger.go:42: 07:40:01 | artifacts | Creating namespace: kuttl-test-poetic-marten logger.go:42: 07:40:01 | artifacts | artifacts events from ns kuttl-test-poetic-marten: logger.go:42: 07:40:01 | artifacts | Deleting namespace: kuttl-test-poetic-marten === CONT kuttl/harness/collector-otlp-production-grpc logger.go:42: 07:40:07 | collector-otlp-production-grpc | Creating namespace: kuttl-test-artistic-mammal logger.go:42: 07:40:07 | collector-otlp-production-grpc/1-install | starting test step 1-install logger.go:42: 07:40:07 | collector-otlp-production-grpc/1-install | Jaeger:kuttl-test-artistic-mammal/my-jaeger created logger.go:42: 07:40:43 | collector-otlp-production-grpc/1-install | test step completed 1-install logger.go:42: 07:40:43 | collector-otlp-production-grpc/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:40:43 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:40:44 | collector-otlp-production-grpc/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:40:50 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:40:51 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:40:51 | collector-otlp-production-grpc/2-smoke-test | job.batch/report-span created logger.go:42: 07:40:51 | collector-otlp-production-grpc/2-smoke-test | job.batch/check-span created logger.go:42: 07:41:10 | collector-otlp-production-grpc/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:41:10 | collector-otlp-production-grpc | collector-otlp-production-grpc events from ns kuttl-test-artistic-mammal: logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:13 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestartisticmammalmyjaeger-1-64599d2vbff Binding Scheduled Successfully assigned kuttl-test-artistic-mammal/elasticsearch-cdm-kuttltestartisticmammalmyjaeger-1-64599d2vbff to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:13 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestartisticmammalmyjaeger-1-64599d2vbff FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:13 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestartisticmammalmyjaeger-1-64599ddd76 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestartisticmammalmyjaeger-1-64599d2vbff replicaset-controller logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:13 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestartisticmammalmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestartisticmammalmyjaeger-1-64599ddd76 to 1 deployment-controller logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestartisticmammalmyjaeger-1-64599d2vbff AddedInterface Add eth0 [10.128.2.62/23] from ovn-kubernetes multus logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestartisticmammalmyjaeger-1-64599d2vbff.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestartisticmammalmyjaeger-1-64599d2vbff.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestartisticmammalmyjaeger-1-64599d2vbff.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestartisticmammalmyjaeger-1-64599d2vbff.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestartisticmammalmyjaeger-1-64599d2vbff.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestartisticmammalmyjaeger-1-64599d2vbff.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:24 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestartisticmammalmyjaeger-1-64599d2vbff.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:29 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestartisticmammalmyjaeger-1-64599d2vbff.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:40 +0000 UTC Normal Pod my-jaeger-collector-bd5b586b-kjnpr Binding Scheduled Successfully assigned kuttl-test-artistic-mammal/my-jaeger-collector-bd5b586b-kjnpr to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:40 +0000 UTC Normal Pod my-jaeger-collector-bd5b586b-kjnpr AddedInterface Add eth0 [10.129.2.73/23] from ovn-kubernetes multus logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:40 +0000 UTC Normal Pod my-jaeger-collector-bd5b586b-kjnpr.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" already present on machine kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:40 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-bd5b586b SuccessfulCreate Created pod: my-jaeger-collector-bd5b586b-kjnpr replicaset-controller logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:40 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-bd5b586b to 1 deployment-controller logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:40 +0000 UTC Normal Pod my-jaeger-query-7c999dbd5f-trxck Binding Scheduled Successfully assigned kuttl-test-artistic-mammal/my-jaeger-query-7c999dbd5f-trxck to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:40 +0000 UTC Normal Pod my-jaeger-query-7c999dbd5f-trxck AddedInterface Add eth0 [10.131.0.66/23] from ovn-kubernetes multus logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:40 +0000 UTC Normal Pod my-jaeger-query-7c999dbd5f-trxck.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:40 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7c999dbd5f SuccessfulCreate Created pod: my-jaeger-query-7c999dbd5f-trxck replicaset-controller logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:40 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7c999dbd5f to 1 deployment-controller logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:41 +0000 UTC Normal Pod my-jaeger-collector-bd5b586b-kjnpr.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:41 +0000 UTC Normal Pod my-jaeger-collector-bd5b586b-kjnpr.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:41 +0000 UTC Normal Pod my-jaeger-query-7c999dbd5f-trxck.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:41 +0000 UTC Normal Pod my-jaeger-query-7c999dbd5f-trxck.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:41 +0000 UTC Normal Pod my-jaeger-query-7c999dbd5f-trxck.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:41 +0000 UTC Normal Pod my-jaeger-query-7c999dbd5f-trxck.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:41 +0000 UTC Normal Pod my-jaeger-query-7c999dbd5f-trxck.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:41 +0000 UTC Normal Pod my-jaeger-query-7c999dbd5f-trxck.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:41 +0000 UTC Normal Pod my-jaeger-query-7c999dbd5f-trxck.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:41 +0000 UTC Normal Pod my-jaeger-query-7c999dbd5f-trxck.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:47 +0000 UTC Normal Pod my-jaeger-query-7c999dbd5f-trxck.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:47 +0000 UTC Normal Pod my-jaeger-query-7c999dbd5f-trxck.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:47 +0000 UTC Normal Pod my-jaeger-query-7c999dbd5f-trxck.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:47 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7c999dbd5f SuccessfulDelete Deleted pod: my-jaeger-query-7c999dbd5f-trxck replicaset-controller logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:47 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-7c999dbd5f to 0 from 1 deployment-controller logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:47 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-65cbcc8d4 to 1 deployment-controller logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:48 +0000 UTC Normal Pod my-jaeger-query-65cbcc8d4-tgl5k Binding Scheduled Successfully assigned kuttl-test-artistic-mammal/my-jaeger-query-65cbcc8d4-tgl5k to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:48 +0000 UTC Normal Pod my-jaeger-query-65cbcc8d4-tgl5k AddedInterface Add eth0 [10.131.0.67/23] from ovn-kubernetes multus logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:48 +0000 UTC Normal Pod my-jaeger-query-65cbcc8d4-tgl5k.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:48 +0000 UTC Normal Pod my-jaeger-query-65cbcc8d4-tgl5k.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:48 +0000 UTC Normal Pod my-jaeger-query-65cbcc8d4-tgl5k.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:48 +0000 UTC Normal Pod my-jaeger-query-65cbcc8d4-tgl5k.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:48 +0000 UTC Normal Pod my-jaeger-query-65cbcc8d4-tgl5k.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:48 +0000 UTC Normal Pod my-jaeger-query-65cbcc8d4-tgl5k.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:48 +0000 UTC Normal Pod my-jaeger-query-65cbcc8d4-tgl5k.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:48 +0000 UTC Normal Pod my-jaeger-query-65cbcc8d4-tgl5k.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:48 +0000 UTC Normal Pod my-jaeger-query-65cbcc8d4-tgl5k.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:48 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-65cbcc8d4 SuccessfulCreate Created pod: my-jaeger-query-65cbcc8d4-tgl5k replicaset-controller logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:51 +0000 UTC Normal Pod check-span-gcv4j Binding Scheduled Successfully assigned kuttl-test-artistic-mammal/check-span-gcv4j to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:51 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-gcv4j job-controller logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:51 +0000 UTC Normal Pod report-span-pr899 Binding Scheduled Successfully assigned kuttl-test-artistic-mammal/report-span-pr899 to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:51 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-pr899 job-controller logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:52 +0000 UTC Normal Pod check-span-gcv4j AddedInterface Add eth0 [10.129.2.75/23] from ovn-kubernetes multus logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:52 +0000 UTC Normal Pod check-span-gcv4j.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:52 +0000 UTC Normal Pod check-span-gcv4j.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:52 +0000 UTC Normal Pod check-span-gcv4j.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:52 +0000 UTC Normal Pod report-span-pr899 AddedInterface Add eth0 [10.129.2.74/23] from ovn-kubernetes multus logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:52 +0000 UTC Normal Pod report-span-pr899.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:52 +0000 UTC Normal Pod report-span-pr899.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:52 +0000 UTC Normal Pod report-span-pr899.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:40:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:41:10 | collector-otlp-production-grpc | 2024-03-04 07:41:10 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:41:10 | collector-otlp-production-grpc | Deleting namespace: kuttl-test-artistic-mammal === CONT kuttl/harness/set-custom-img logger.go:42: 07:41:22 | set-custom-img | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:41:22 | set-custom-img | Ignoring check-collector-img.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:41:22 | set-custom-img | Creating namespace: kuttl-test-vital-possum logger.go:42: 07:41:22 | set-custom-img/1-install | starting test step 1-install logger.go:42: 07:41:22 | set-custom-img/1-install | Jaeger:kuttl-test-vital-possum/my-jaeger created logger.go:42: 07:41:58 | set-custom-img/1-install | test step completed 1-install logger.go:42: 07:41:58 | set-custom-img/2-install | starting test step 2-install logger.go:42: 07:41:58 | set-custom-img/2-install | Jaeger:kuttl-test-vital-possum/my-jaeger updated logger.go:42: 07:41:58 | set-custom-img/2-install | test step completed 2-install logger.go:42: 07:41:58 | set-custom-img/3-check-image | starting test step 3-check-image logger.go:42: 07:41:58 | set-custom-img/3-check-image | running command: [sh -c ./check-collector-img.sh] logger.go:42: 07:41:58 | set-custom-img/3-check-image | Collector image missmatch. Expected: test. Has: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9 logger.go:42: 07:42:03 | set-custom-img/3-check-image | Collector image asserted properly! logger.go:42: 07:42:03 | set-custom-img/3-check-image | test step completed 3-check-image logger.go:42: 07:42:03 | set-custom-img | set-custom-img events from ns kuttl-test-vital-possum: logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:28 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestvitalpossummyjaeger-1-6c559db8dd SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestvitalpossummyjaeger-1-6c559db8dnh4pj replicaset-controller logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestvitalpossummyjaeger-1-6c559db8dnh4pj Binding Scheduled Successfully assigned kuttl-test-vital-possum/elasticsearch-cdm-kuttltestvitalpossummyjaeger-1-6c559db8dnh4pj to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:28 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestvitalpossummyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestvitalpossummyjaeger-1-6c559db8dd to 1 deployment-controller logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestvitalpossummyjaeger-1-6c559db8dnh4pj AddedInterface Add eth0 [10.128.2.63/23] from ovn-kubernetes multus logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestvitalpossummyjaeger-1-6c559db8dnh4pj.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestvitalpossummyjaeger-1-6c559db8dnh4pj.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestvitalpossummyjaeger-1-6c559db8dnh4pj.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestvitalpossummyjaeger-1-6c559db8dnh4pj.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestvitalpossummyjaeger-1-6c559db8dnh4pj.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestvitalpossummyjaeger-1-6c559db8dnh4pj.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:39 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestvitalpossummyjaeger-1-6c559db8dnh4pj.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:44 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestvitalpossummyjaeger-1-6c559db8dnh4pj.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:55 +0000 UTC Normal Pod my-jaeger-collector-758855d5c5-b9wtq Binding Scheduled Successfully assigned kuttl-test-vital-possum/my-jaeger-collector-758855d5c5-b9wtq to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:55 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-758855d5c5 SuccessfulCreate Created pod: my-jaeger-collector-758855d5c5-b9wtq replicaset-controller logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:55 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-758855d5c5 to 1 deployment-controller logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:55 +0000 UTC Normal Pod my-jaeger-query-567b8d5486-xpq6p Binding Scheduled Successfully assigned kuttl-test-vital-possum/my-jaeger-query-567b8d5486-xpq6p to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:55 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-567b8d5486 SuccessfulCreate Created pod: my-jaeger-query-567b8d5486-xpq6p replicaset-controller logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:55 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-567b8d5486 to 1 deployment-controller logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:56 +0000 UTC Normal Pod my-jaeger-collector-758855d5c5-b9wtq AddedInterface Add eth0 [10.129.2.76/23] from ovn-kubernetes multus logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:56 +0000 UTC Normal Pod my-jaeger-collector-758855d5c5-b9wtq.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" already present on machine kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:56 +0000 UTC Normal Pod my-jaeger-collector-758855d5c5-b9wtq.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:56 +0000 UTC Normal Pod my-jaeger-collector-758855d5c5-b9wtq.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:56 +0000 UTC Normal Pod my-jaeger-query-567b8d5486-xpq6p AddedInterface Add eth0 [10.131.0.68/23] from ovn-kubernetes multus logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:56 +0000 UTC Normal Pod my-jaeger-query-567b8d5486-xpq6p.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:56 +0000 UTC Normal Pod my-jaeger-query-567b8d5486-xpq6p.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:56 +0000 UTC Normal Pod my-jaeger-query-567b8d5486-xpq6p.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:56 +0000 UTC Normal Pod my-jaeger-query-567b8d5486-xpq6p.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:56 +0000 UTC Normal Pod my-jaeger-query-567b8d5486-xpq6p.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:56 +0000 UTC Normal Pod my-jaeger-query-567b8d5486-xpq6p.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:56 +0000 UTC Normal Pod my-jaeger-query-567b8d5486-xpq6p.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:57 +0000 UTC Normal Pod my-jaeger-query-567b8d5486-xpq6p.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:41:57 +0000 UTC Normal Pod my-jaeger-query-567b8d5486-xpq6p.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:42:01 +0000 UTC Normal Pod my-jaeger-collector-758855d5c5-b9wtq.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:42:01 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-758855d5c5 SuccessfulDelete Deleted pod: my-jaeger-collector-758855d5c5-b9wtq replicaset-controller logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:42:01 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-758855d5c5 to 0 from 1 deployment-controller logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:42:02 +0000 UTC Normal Pod my-jaeger-collector-64dfdc9d99-5drj2 Binding Scheduled Successfully assigned kuttl-test-vital-possum/my-jaeger-collector-64dfdc9d99-5drj2 to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:42:02 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-64dfdc9d99 SuccessfulCreate Created pod: my-jaeger-collector-64dfdc9d99-5drj2 replicaset-controller logger.go:42: 07:42:03 | set-custom-img | 2024-03-04 07:42:02 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-64dfdc9d99 to 1 deployment-controller logger.go:42: 07:42:03 | set-custom-img | Deleting namespace: kuttl-test-vital-possum === CONT kuttl/harness/collector-otlp-production-http logger.go:42: 07:42:10 | collector-otlp-production-http | Creating namespace: kuttl-test-wealthy-anteater logger.go:42: 07:42:10 | collector-otlp-production-http/1-install | starting test step 1-install logger.go:42: 07:42:10 | collector-otlp-production-http/1-install | Jaeger:kuttl-test-wealthy-anteater/my-jaeger created logger.go:42: 07:42:46 | collector-otlp-production-http/1-install | test step completed 1-install logger.go:42: 07:42:46 | collector-otlp-production-http/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:42:46 | collector-otlp-production-http/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:42:48 | collector-otlp-production-http/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:42:54 | collector-otlp-production-http/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:42:54 | collector-otlp-production-http/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:42:55 | collector-otlp-production-http/2-smoke-test | job.batch/report-span created logger.go:42: 07:42:55 | collector-otlp-production-http/2-smoke-test | job.batch/check-span created logger.go:42: 07:43:07 | collector-otlp-production-http/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:43:07 | collector-otlp-production-http | collector-otlp-production-http events from ns kuttl-test-wealthy-anteater: logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:16 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestwealthyanteatermyjaeger-1-66588b8c4b SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestwealthyanteatermyjaeger-1-66588kk2pw replicaset-controller logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwealthyanteatermyjaeger-1-66588kk2pw Binding Scheduled Successfully assigned kuttl-test-wealthy-anteater/elasticsearch-cdm-kuttltestwealthyanteatermyjaeger-1-66588kk2pw to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:16 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestwealthyanteatermyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestwealthyanteatermyjaeger-1-66588b8c4b to 1 deployment-controller logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwealthyanteatermyjaeger-1-66588kk2pw AddedInterface Add eth0 [10.128.2.64/23] from ovn-kubernetes multus logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwealthyanteatermyjaeger-1-66588kk2pw.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwealthyanteatermyjaeger-1-66588kk2pw.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwealthyanteatermyjaeger-1-66588kk2pw.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwealthyanteatermyjaeger-1-66588kk2pw.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwealthyanteatermyjaeger-1-66588kk2pw.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwealthyanteatermyjaeger-1-66588kk2pw.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:27 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestwealthyanteatermyjaeger-1-66588kk2pw.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:32 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestwealthyanteatermyjaeger-1-66588kk2pw.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:43 +0000 UTC Normal Pod my-jaeger-collector-6758764d8d-29r48 Binding Scheduled Successfully assigned kuttl-test-wealthy-anteater/my-jaeger-collector-6758764d8d-29r48 to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:43 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-6758764d8d SuccessfulCreate Created pod: my-jaeger-collector-6758764d8d-29r48 replicaset-controller logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:43 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-6758764d8d to 1 deployment-controller logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:43 +0000 UTC Normal Pod my-jaeger-query-7b6d459467-j6bpq Binding Scheduled Successfully assigned kuttl-test-wealthy-anteater/my-jaeger-query-7b6d459467-j6bpq to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:43 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7b6d459467 SuccessfulCreate Created pod: my-jaeger-query-7b6d459467-j6bpq replicaset-controller logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:43 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7b6d459467 to 1 deployment-controller logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:44 +0000 UTC Normal Pod my-jaeger-collector-6758764d8d-29r48 AddedInterface Add eth0 [10.129.2.78/23] from ovn-kubernetes multus logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:44 +0000 UTC Normal Pod my-jaeger-collector-6758764d8d-29r48.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" already present on machine kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:44 +0000 UTC Normal Pod my-jaeger-collector-6758764d8d-29r48.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:44 +0000 UTC Normal Pod my-jaeger-collector-6758764d8d-29r48.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:44 +0000 UTC Normal Pod my-jaeger-query-7b6d459467-j6bpq AddedInterface Add eth0 [10.131.0.69/23] from ovn-kubernetes multus logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:44 +0000 UTC Normal Pod my-jaeger-query-7b6d459467-j6bpq.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:44 +0000 UTC Normal Pod my-jaeger-query-7b6d459467-j6bpq.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:44 +0000 UTC Normal Pod my-jaeger-query-7b6d459467-j6bpq.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:44 +0000 UTC Normal Pod my-jaeger-query-7b6d459467-j6bpq.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:44 +0000 UTC Normal Pod my-jaeger-query-7b6d459467-j6bpq.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:44 +0000 UTC Normal Pod my-jaeger-query-7b6d459467-j6bpq.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:44 +0000 UTC Normal Pod my-jaeger-query-7b6d459467-j6bpq.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:44 +0000 UTC Normal Pod my-jaeger-query-7b6d459467-j6bpq.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:44 +0000 UTC Normal Pod my-jaeger-query-7b6d459467-j6bpq.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:49 +0000 UTC Normal Pod my-jaeger-query-7b6d459467-j6bpq.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:49 +0000 UTC Normal Pod my-jaeger-query-7b6d459467-j6bpq.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:49 +0000 UTC Normal Pod my-jaeger-query-7b6d459467-j6bpq.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:49 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7b6d459467 SuccessfulDelete Deleted pod: my-jaeger-query-7b6d459467-j6bpq replicaset-controller logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:49 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-7b6d459467 to 0 from 1 deployment-controller logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:50 +0000 UTC Normal Pod my-jaeger-query-6bdfc8867d-z7qx7 Binding Scheduled Successfully assigned kuttl-test-wealthy-anteater/my-jaeger-query-6bdfc8867d-z7qx7 to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:50 +0000 UTC Normal Pod my-jaeger-query-6bdfc8867d-z7qx7 AddedInterface Add eth0 [10.131.0.70/23] from ovn-kubernetes multus logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:50 +0000 UTC Normal Pod my-jaeger-query-6bdfc8867d-z7qx7.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:50 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-6bdfc8867d SuccessfulCreate Created pod: my-jaeger-query-6bdfc8867d-z7qx7 replicaset-controller logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:50 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-6bdfc8867d to 1 deployment-controller logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:51 +0000 UTC Normal Pod my-jaeger-query-6bdfc8867d-z7qx7.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:51 +0000 UTC Normal Pod my-jaeger-query-6bdfc8867d-z7qx7.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:51 +0000 UTC Normal Pod my-jaeger-query-6bdfc8867d-z7qx7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:51 +0000 UTC Normal Pod my-jaeger-query-6bdfc8867d-z7qx7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:51 +0000 UTC Normal Pod my-jaeger-query-6bdfc8867d-z7qx7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:51 +0000 UTC Normal Pod my-jaeger-query-6bdfc8867d-z7qx7.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:51 +0000 UTC Normal Pod my-jaeger-query-6bdfc8867d-z7qx7.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:51 +0000 UTC Normal Pod my-jaeger-query-6bdfc8867d-z7qx7.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:55 +0000 UTC Normal Pod check-span-24lf5 Binding Scheduled Successfully assigned kuttl-test-wealthy-anteater/check-span-24lf5 to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:55 +0000 UTC Normal Pod check-span-24lf5 AddedInterface Add eth0 [10.129.2.80/23] from ovn-kubernetes multus logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:55 +0000 UTC Normal Pod check-span-24lf5.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:55 +0000 UTC Normal Pod check-span-24lf5.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:55 +0000 UTC Normal Pod check-span-24lf5.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:55 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-24lf5 job-controller logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:55 +0000 UTC Normal Pod report-span-rp7ln Binding Scheduled Successfully assigned kuttl-test-wealthy-anteater/report-span-rp7ln to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:55 +0000 UTC Normal Pod report-span-rp7ln AddedInterface Add eth0 [10.129.2.79/23] from ovn-kubernetes multus logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:55 +0000 UTC Normal Pod report-span-rp7ln.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:55 +0000 UTC Normal Pod report-span-rp7ln.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:55 +0000 UTC Normal Pod report-span-rp7ln.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:55 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-rp7ln job-controller logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:42:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:43:07 | collector-otlp-production-http | 2024-03-04 07:43:07 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:43:07 | collector-otlp-production-http | Deleting namespace: kuttl-test-wealthy-anteater === CONT kuttl/harness/collector-otlp-allinone-grpc logger.go:42: 07:43:19 | collector-otlp-allinone-grpc | Creating namespace: kuttl-test-holy-roughy logger.go:42: 07:43:19 | collector-otlp-allinone-grpc/0-install | starting test step 0-install logger.go:42: 07:43:19 | collector-otlp-allinone-grpc/0-install | Jaeger:kuttl-test-holy-roughy/my-jaeger created logger.go:42: 07:43:25 | collector-otlp-allinone-grpc/0-install | test step completed 0-install logger.go:42: 07:43:25 | collector-otlp-allinone-grpc/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:43:25 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:43:26 | collector-otlp-allinone-grpc/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:43:33 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:43:33 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:43:34 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/report-span created logger.go:42: 07:43:34 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/check-span created logger.go:42: 07:43:53 | collector-otlp-allinone-grpc/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | collector-otlp-allinone-grpc events from ns kuttl-test-holy-roughy: logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:22 +0000 UTC Normal Pod my-jaeger-547488f4cc-mlrvg Binding Scheduled Successfully assigned kuttl-test-holy-roughy/my-jaeger-547488f4cc-mlrvg to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:22 +0000 UTC Normal ReplicaSet.apps my-jaeger-547488f4cc SuccessfulCreate Created pod: my-jaeger-547488f4cc-mlrvg replicaset-controller logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:22 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-547488f4cc to 1 deployment-controller logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:23 +0000 UTC Normal Pod my-jaeger-547488f4cc-mlrvg AddedInterface Add eth0 [10.128.2.65/23] from ovn-kubernetes multus logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:23 +0000 UTC Normal Pod my-jaeger-547488f4cc-mlrvg.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:23 +0000 UTC Normal Pod my-jaeger-547488f4cc-mlrvg.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:23 +0000 UTC Normal Pod my-jaeger-547488f4cc-mlrvg.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:23 +0000 UTC Normal Pod my-jaeger-547488f4cc-mlrvg.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:23 +0000 UTC Normal Pod my-jaeger-547488f4cc-mlrvg.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:23 +0000 UTC Normal Pod my-jaeger-547488f4cc-mlrvg.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:30 +0000 UTC Normal Pod my-jaeger-547488f4cc-mlrvg.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:30 +0000 UTC Normal Pod my-jaeger-547488f4cc-mlrvg.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:30 +0000 UTC Normal ReplicaSet.apps my-jaeger-547488f4cc SuccessfulDelete Deleted pod: my-jaeger-547488f4cc-mlrvg replicaset-controller logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:30 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-547488f4cc to 0 from 1 deployment-controller logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:31 +0000 UTC Normal Pod my-jaeger-54ff765dfc-jlc62 Binding Scheduled Successfully assigned kuttl-test-holy-roughy/my-jaeger-54ff765dfc-jlc62 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:31 +0000 UTC Normal ReplicaSet.apps my-jaeger-54ff765dfc SuccessfulCreate Created pod: my-jaeger-54ff765dfc-jlc62 replicaset-controller logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:31 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-54ff765dfc to 1 deployment-controller logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:32 +0000 UTC Normal Pod my-jaeger-54ff765dfc-jlc62 AddedInterface Add eth0 [10.128.2.66/23] from ovn-kubernetes multus logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:32 +0000 UTC Normal Pod my-jaeger-54ff765dfc-jlc62.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:32 +0000 UTC Normal Pod my-jaeger-54ff765dfc-jlc62.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:32 +0000 UTC Normal Pod my-jaeger-54ff765dfc-jlc62.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:32 +0000 UTC Normal Pod my-jaeger-54ff765dfc-jlc62.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:32 +0000 UTC Normal Pod my-jaeger-54ff765dfc-jlc62.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:32 +0000 UTC Normal Pod my-jaeger-54ff765dfc-jlc62.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:34 +0000 UTC Normal Pod check-span-j5knn Binding Scheduled Successfully assigned kuttl-test-holy-roughy/check-span-j5knn to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:34 +0000 UTC Normal Pod check-span-j5knn AddedInterface Add eth0 [10.131.0.71/23] from ovn-kubernetes multus logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:34 +0000 UTC Normal Pod check-span-j5knn.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:34 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-j5knn job-controller logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:34 +0000 UTC Normal Pod report-span-rr4bd Binding Scheduled Successfully assigned kuttl-test-holy-roughy/report-span-rr4bd to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:34 +0000 UTC Normal Pod report-span-rr4bd AddedInterface Add eth0 [10.129.2.81/23] from ovn-kubernetes multus logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:34 +0000 UTC Normal Pod report-span-rr4bd.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:34 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-rr4bd job-controller logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:35 +0000 UTC Normal Pod check-span-j5knn.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:35 +0000 UTC Normal Pod check-span-j5knn.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:35 +0000 UTC Normal Pod report-span-rr4bd.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:35 +0000 UTC Normal Pod report-span-rr4bd.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | 2024-03-04 07:43:53 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:43:53 | collector-otlp-allinone-grpc | Deleting namespace: kuttl-test-holy-roughy === CONT kuttl/harness/collector-otlp-allinone-http logger.go:42: 07:44:06 | collector-otlp-allinone-http | Creating namespace: kuttl-test-intense-rat logger.go:42: 07:44:06 | collector-otlp-allinone-http/0-install | starting test step 0-install logger.go:42: 07:44:06 | collector-otlp-allinone-http/0-install | Jaeger:kuttl-test-intense-rat/my-jaeger created logger.go:42: 07:44:12 | collector-otlp-allinone-http/0-install | test step completed 0-install logger.go:42: 07:44:12 | collector-otlp-allinone-http/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:44:12 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:44:13 | collector-otlp-allinone-http/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:44:19 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:44:20 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:44:20 | collector-otlp-allinone-http/1-smoke-test | job.batch/report-span created logger.go:42: 07:44:20 | collector-otlp-allinone-http/1-smoke-test | job.batch/check-span created logger.go:42: 07:44:32 | collector-otlp-allinone-http/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:44:32 | collector-otlp-allinone-http | collector-otlp-allinone-http events from ns kuttl-test-intense-rat: logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:09 +0000 UTC Normal Pod my-jaeger-5947b58bfb-2sf66 Binding Scheduled Successfully assigned kuttl-test-intense-rat/my-jaeger-5947b58bfb-2sf66 to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:09 +0000 UTC Normal ReplicaSet.apps my-jaeger-5947b58bfb SuccessfulCreate Created pod: my-jaeger-5947b58bfb-2sf66 replicaset-controller logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:09 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5947b58bfb to 1 deployment-controller logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:10 +0000 UTC Normal Pod my-jaeger-5947b58bfb-2sf66 AddedInterface Add eth0 [10.128.2.67/23] from ovn-kubernetes multus logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:10 +0000 UTC Normal Pod my-jaeger-5947b58bfb-2sf66.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:10 +0000 UTC Normal Pod my-jaeger-5947b58bfb-2sf66.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:10 +0000 UTC Normal Pod my-jaeger-5947b58bfb-2sf66.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:10 +0000 UTC Normal Pod my-jaeger-5947b58bfb-2sf66.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:11 +0000 UTC Normal Pod my-jaeger-5947b58bfb-2sf66.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:11 +0000 UTC Normal Pod my-jaeger-5947b58bfb-2sf66.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:14 +0000 UTC Normal Pod my-jaeger-5947b58bfb-2sf66.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:14 +0000 UTC Normal Pod my-jaeger-5947b58bfb-2sf66.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:14 +0000 UTC Normal ReplicaSet.apps my-jaeger-5947b58bfb SuccessfulDelete Deleted pod: my-jaeger-5947b58bfb-2sf66 replicaset-controller logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:14 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-5947b58bfb to 0 from 1 deployment-controller logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:15 +0000 UTC Normal Pod my-jaeger-756455bff-z97sk Binding Scheduled Successfully assigned kuttl-test-intense-rat/my-jaeger-756455bff-z97sk to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:15 +0000 UTC Normal ReplicaSet.apps my-jaeger-756455bff SuccessfulCreate Created pod: my-jaeger-756455bff-z97sk replicaset-controller logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:15 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-756455bff to 1 deployment-controller logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:16 +0000 UTC Normal Pod my-jaeger-756455bff-z97sk AddedInterface Add eth0 [10.128.2.68/23] from ovn-kubernetes multus logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:16 +0000 UTC Normal Pod my-jaeger-756455bff-z97sk.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:5050eb689ff1ab014680f406f3abdf8a796e582684df9372d4ad00cb6c401c66" already present on machine kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:16 +0000 UTC Normal Pod my-jaeger-756455bff-z97sk.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:16 +0000 UTC Normal Pod my-jaeger-756455bff-z97sk.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:16 +0000 UTC Normal Pod my-jaeger-756455bff-z97sk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:16 +0000 UTC Normal Pod my-jaeger-756455bff-z97sk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:16 +0000 UTC Normal Pod my-jaeger-756455bff-z97sk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:20 +0000 UTC Normal Pod check-span-q4vjn Binding Scheduled Successfully assigned kuttl-test-intense-rat/check-span-q4vjn to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:20 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-q4vjn job-controller logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:20 +0000 UTC Normal Pod report-span-57d5x Binding Scheduled Successfully assigned kuttl-test-intense-rat/report-span-57d5x to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:20 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-57d5x job-controller logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:21 +0000 UTC Normal Pod check-span-q4vjn AddedInterface Add eth0 [10.131.0.72/23] from ovn-kubernetes multus logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:21 +0000 UTC Normal Pod check-span-q4vjn.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:21 +0000 UTC Normal Pod check-span-q4vjn.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:21 +0000 UTC Normal Pod check-span-q4vjn.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:21 +0000 UTC Normal Pod report-span-57d5x AddedInterface Add eth0 [10.129.2.82/23] from ovn-kubernetes multus logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:21 +0000 UTC Normal Pod report-span-57d5x.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-yjw0rx4s/pipeline@sha256:05e69671b50e8b4f9b2c26b043934bc536c228e83a685b152a1fa4fb13f66d5f" already present on machine kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:21 +0000 UTC Normal Pod report-span-57d5x.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:21 +0000 UTC Normal Pod report-span-57d5x.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:44:32 | collector-otlp-allinone-http | 2024-03-04 07:44:32 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:44:32 | collector-otlp-allinone-http | Deleting namespace: kuttl-test-intense-rat === CONT kuttl/harness/collector-autoscale logger.go:42: 07:44:45 | collector-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:44:45 | collector-autoscale | Creating namespace: kuttl-test-upward-gnat logger.go:42: 07:44:45 | collector-autoscale/1-install | starting test step 1-install logger.go:42: 07:44:45 | collector-autoscale/1-install | Jaeger:kuttl-test-upward-gnat/simple-prod created logger.go:42: 07:45:20 | collector-autoscale/1-install | test step completed 1-install logger.go:42: 07:45:20 | collector-autoscale/2- | starting test step 2- logger.go:42: 07:45:20 | collector-autoscale/2- | test step completed 2- logger.go:42: 07:45:20 | collector-autoscale | collector-autoscale events from ns kuttl-test-upward-gnat: logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:44:50 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestupwardgnatsimpleprod-1-78fbccc568 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestupwardgnatsimpleprod-1-78fbccc5zgn5r replicaset-controller logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:44:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupwardgnatsimpleprod-1-78fbccc5zgn5r Binding Scheduled Successfully assigned kuttl-test-upward-gnat/elasticsearch-cdm-kuttltestupwardgnatsimpleprod-1-78fbccc5zgn5r to ip-10-0-15-92.ec2.internal default-scheduler logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:44:50 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestupwardgnatsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestupwardgnatsimpleprod-1-78fbccc568 to 1 deployment-controller logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:44:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupwardgnatsimpleprod-1-78fbccc5zgn5r AddedInterface Add eth0 [10.128.2.69/23] from ovn-kubernetes multus logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:44:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupwardgnatsimpleprod-1-78fbccc5zgn5r.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:44:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupwardgnatsimpleprod-1-78fbccc5zgn5r.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:44:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupwardgnatsimpleprod-1-78fbccc5zgn5r.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:44:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupwardgnatsimpleprod-1-78fbccc5zgn5r.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:44:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupwardgnatsimpleprod-1-78fbccc5zgn5r.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:44:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupwardgnatsimpleprod-1-78fbccc5zgn5r.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:01 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestupwardgnatsimpleprod-1-78fbccc5zgn5r.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:06 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestupwardgnatsimpleprod-1-78fbccc5zgn5r.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:17 +0000 UTC Normal Pod simple-prod-collector-85495f76dc-4f5jk Binding Scheduled Successfully assigned kuttl-test-upward-gnat/simple-prod-collector-85495f76dc-4f5jk to ip-10-0-106-109.ec2.internal default-scheduler logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:17 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-85495f76dc SuccessfulCreate Created pod: simple-prod-collector-85495f76dc-4f5jk replicaset-controller logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:17 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-85495f76dc to 1 deployment-controller logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:17 +0000 UTC Normal Pod simple-prod-query-67d6bf4d49-k5k8p Binding Scheduled Successfully assigned kuttl-test-upward-gnat/simple-prod-query-67d6bf4d49-k5k8p to ip-10-0-61-248.ec2.internal default-scheduler logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:17 +0000 UTC Normal ReplicaSet.apps simple-prod-query-67d6bf4d49 SuccessfulCreate Created pod: simple-prod-query-67d6bf4d49-k5k8p replicaset-controller logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:17 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-67d6bf4d49 to 1 deployment-controller logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:18 +0000 UTC Normal Pod simple-prod-collector-85495f76dc-4f5jk AddedInterface Add eth0 [10.131.0.73/23] from ovn-kubernetes multus logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:18 +0000 UTC Normal Pod simple-prod-collector-85495f76dc-4f5jk.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:2f9622a3e5b7eacd10634cb3cb66023a299801e9b73a16e93b1f9a3459211cd9" already present on machine kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:18 +0000 UTC Normal Pod simple-prod-collector-85495f76dc-4f5jk.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:18 +0000 UTC Normal Pod simple-prod-collector-85495f76dc-4f5jk.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:18 +0000 UTC Normal Pod simple-prod-query-67d6bf4d49-k5k8p AddedInterface Add eth0 [10.129.2.84/23] from ovn-kubernetes multus logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:18 +0000 UTC Normal Pod simple-prod-query-67d6bf4d49-k5k8p.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:ded77217864e9502de4215ef08bd40b047ffe35487bc449e42d086dcbcc1ac3a" already present on machine kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:18 +0000 UTC Normal Pod simple-prod-query-67d6bf4d49-k5k8p.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:18 +0000 UTC Normal Pod simple-prod-query-67d6bf4d49-k5k8p.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:18 +0000 UTC Normal Pod simple-prod-query-67d6bf4d49-k5k8p.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:18 +0000 UTC Normal Pod simple-prod-query-67d6bf4d49-k5k8p.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:18 +0000 UTC Normal Pod simple-prod-query-67d6bf4d49-k5k8p.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:18 +0000 UTC Normal Pod simple-prod-query-67d6bf4d49-k5k8p.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:de5b98d3e5cef557d3e73afddfca3e7db5b2901d71bd58b3ec4d9d58a2e58b51" already present on machine kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:18 +0000 UTC Normal Pod simple-prod-query-67d6bf4d49-k5k8p.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:45:20 | collector-autoscale | 2024-03-04 07:45:18 +0000 UTC Normal Pod simple-prod-query-67d6bf4d49-k5k8p.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:45:20 | collector-autoscale | Deleting namespace: kuttl-test-upward-gnat === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (324.73s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.87s) --- PASS: kuttl/harness/collector-otlp-production-grpc (75.10s) --- PASS: kuttl/harness/set-custom-img (47.18s) --- PASS: kuttl/harness/collector-otlp-production-http (69.14s) --- PASS: kuttl/harness/collector-otlp-allinone-grpc (46.81s) --- PASS: kuttl/harness/collector-otlp-allinone-http (39.04s) --- PASS: kuttl/harness/collector-autoscale (41.55s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name miscellaneous --report --output /logs/artifacts/miscellaneous.xml ./artifacts/kuttl-report.xml time="2024-03-04T07:45:26Z" level=debug msg="Setting a new name for the test suites" time="2024-03-04T07:45:26Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-03-04T07:45:26Z" level=debug msg="normalizing test case names" time="2024-03-04T07:45:26Z" level=debug msg="miscellaneous/artifacts -> miscellaneous_artifacts" time="2024-03-04T07:45:26Z" level=debug msg="miscellaneous/collector-otlp-production-grpc -> miscellaneous_collector_otlp_production_grpc" time="2024-03-04T07:45:26Z" level=debug msg="miscellaneous/set-custom-img -> miscellaneous_set_custom_img" time="2024-03-04T07:45:26Z" level=debug msg="miscellaneous/collector-otlp-production-http -> miscellaneous_collector_otlp_production_http" time="2024-03-04T07:45:26Z" level=debug msg="miscellaneous/collector-otlp-allinone-grpc -> miscellaneous_collector_otlp_allinone_grpc" time="2024-03-04T07:45:26Z" level=debug msg="miscellaneous/collector-otlp-allinone-http -> miscellaneous_collector_otlp_allinone_http" time="2024-03-04T07:45:26Z" level=debug msg="miscellaneous/collector-autoscale -> miscellaneous_collector_autoscale" +----------------------------------------------+--------+ | NAME | RESULT | +----------------------------------------------+--------+ | miscellaneous_artifacts | passed | | miscellaneous_collector_otlp_production_grpc | passed | | miscellaneous_set_custom_img | passed | | miscellaneous_collector_otlp_production_http | passed | | miscellaneous_collector_otlp_allinone_grpc | passed | | miscellaneous_collector_otlp_allinone_http | passed | | miscellaneous_collector_autoscale | passed | +----------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests'