% Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 3831 100 3831 0 0 28338 0 --:--:-- --:--:-- --:--:-- 28377 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 1953 100 1953 0 0 18379 0 --:--:-- --:--:-- --:--:-- 18424 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 350 100 350 0 0 3067 0 --:--:-- --:--:-- --:--:-- 3097 Installing kuttl Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/kubectl-kuttl https://github.com/kudobuilder/kuttl/releases/download/v0.15.0/kubectl-kuttl_0.15.0_linux_x86_64 KUBECONFIG file is: /tmp/kubeconfig-1441553849 for suite in elasticsearch examples generate miscellaneous sidecar streaming ui upgrade; do \ make run-e2e-tests-$suite ; \ done make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh elasticsearch false true + '[' 3 -ne 3 ']' + test_suite_name=elasticsearch + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/elasticsearch.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-elasticsearch make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true \ KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ ./tests/e2e/elasticsearch/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-12-22-165425 True False 6m58s Cluster version is 4.15.0-0.nightly-2023-12-22-165425' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-12-22-165425 True False 6m58s Cluster version is 4.15.0-0.nightly-2023-12-22-165425' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/elasticsearch/render.sh ++ export SUITE_DIR=./tests/e2e/elasticsearch ++ SUITE_DIR=./tests/e2e/elasticsearch ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/elasticsearch ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + start_test es-from-aio-to-production + '[' 1 -ne 1 ']' + test_name=es-from-aio-to-production + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-from-aio-to-production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-from-aio-to-production\e[0m' Rendering files for test es-from-aio-to-production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-from-aio-to-production + cd es-from-aio-to-production + jaeger_name=my-jaeger + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 03 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=03 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./03-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch.redundancyPolicy="ZeroRedundancy"' ./03-install.yaml + render_smoke_test my-jaeger true 04 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=04 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./04-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test es-increasing-replicas + '[' 1 -ne 1 ']' + test_name=es-increasing-replicas + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-increasing-replicas' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-increasing-replicas\e[0m' Rendering files for test es-increasing-replicas + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-from-aio-to-production + '[' es-from-aio-to-production '!=' _build ']' + cd .. + mkdir -p es-increasing-replicas + cd es-increasing-replicas + jaeger_name=simple-prod + '[' true = true ']' + jaeger_deployment_mode=production_autoprovisioned + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.replicas=2 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.query.replicas=2 ./02-install.yaml + cp ./01-assert.yaml ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.replicas=2 ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .status.readyReplicas=2 ./02-assert.yaml + render_smoke_test simple-prod true 03 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=03 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./03-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + cp ./02-install.yaml ./04-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.elasticsearch.nodeCount=2 ./04-install.yaml + /tmp/jaeger-tests/bin/gomplate -f ./openshift-check-es-nodes.yaml.template -o ./05-check-es-nodes.yaml + '[' true = true ']' + skip_test es-index-cleaner-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-increasing-replicas + '[' es-increasing-replicas '!=' _build ']' + cd .. + rm -rf es-index-cleaner-upstream + warning 'es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_index_cleaner -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-index-cleaner-autoprov + '[' 1 -ne 1 ']' + test_name=es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-index-cleaner-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-index-cleaner-autoprov\e[0m' Rendering files for test es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-index-cleaner-autoprov + cd es-index-cleaner-autoprov + jaeger_name=test-es-index-cleaner-with-prefix + cronjob_name=test-es-index-cleaner-with-prefix-es-index-cleaner + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + cp ../../es-index-cleaner-upstream/04-assert.yaml ../../es-index-cleaner-upstream/README.md . + render_install_jaeger test-es-index-cleaner-with-prefix production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options.es.index-prefix=""' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.enabled=false ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.numberOfDays=0 ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.esIndexCleaner.schedule="*/1 * * * *"' ./01-install.yaml + render_report_spans test-es-index-cleaner-with-prefix true 5 00 true 02 + '[' 6 -ne 6 ']' + jaeger=test-es-index-cleaner-with-prefix + is_secured=true + number_of_spans=5 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + export JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=5 + DAYS=5 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + sed 's~enabled: false~enabled: true~gi' ./01-install.yaml + CRONJOB_NAME=test-es-index-cleaner-with-prefix-es-index-cleaner + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./04-wait-es-index-cleaner.yaml + /tmp/jaeger-tests/bin/gomplate -f ./01-install.yaml -o ./05-install.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 00 06 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=00 + test_step=06 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=test-es-index-cleaner-with-prefix-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=test-es-index-cleaner-with-prefix-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./06-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./06-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.1"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.1"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.1"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.1 ++ version_ge 5.8.1 5.4 +++ echo 5.8.1 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.1 == 5.8.1 + '[' -n '' ']' + skip_test es-index-cleaner-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-index-cleaner-autoprov + '[' es-index-cleaner-autoprov '!=' _build ']' + cd .. + rm -rf es-index-cleaner-managed + warning 'es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + start_test es-multiinstance + '[' 1 -ne 1 ']' + test_name=es-multiinstance + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-multiinstance' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-multiinstance\e[0m' Rendering files for test es-multiinstance + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-multiinstance + cd es-multiinstance + jaeger_name=instance-1 + render_install_jaeger instance-1 production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=instance-1 + JAEGER_NAME=instance-1 + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f ./03-create-second-instance.yaml.template -o 03-create-second-instance.yaml + '[' true = true ']' + skip_test es-rollover-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-rollover-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-multiinstance + '[' es-multiinstance '!=' _build ']' + cd .. + rm -rf es-rollover-upstream + warning 'es-rollover-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_rollover -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-rollover-autoprov + '[' 1 -ne 1 ']' + test_name=es-rollover-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-rollover-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-rollover-autoprov\e[0m' Rendering files for test es-rollover-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-rollover-autoprov + cd es-rollover-autoprov + cp ../../es-rollover-upstream/05-assert.yaml ../../es-rollover-upstream/05-install.yaml ../../es-rollover-upstream/README.md . + jaeger_name=my-jaeger + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_report_spans my-jaeger true 2 00 true 02 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 00 03 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=00 + test_step=03 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./03-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./03-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 01 04 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=01 + test_step=04 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=01 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./04-check-indices.yaml + JOB_NUMBER=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./04-assert.yaml + render_report_spans my-jaeger true 2 02 true 06 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=02 + ensure_reported_spans=true + test_step=06 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=02 + JOB_NUMBER=02 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./06-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./06-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 02 07 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=02 + test_step=07 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=02 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./07-check-indices.yaml + JOB_NUMBER=02 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./07-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' 03 08 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + job_number=03 + test_step=08 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=03 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./08-check-indices.yaml + JOB_NUMBER=03 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./08-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' 04 09 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + job_number=04 + test_step=09 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=04 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./09-check-indices.yaml + JOB_NUMBER=04 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./09-assert.yaml + render_report_spans my-jaeger true 2 03 true 10 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=03 + ensure_reported_spans=true + test_step=10 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=03 + JOB_NUMBER=03 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./10-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./10-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + CRONJOB_NAME=my-jaeger-es-rollover + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./11-wait-rollover.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-000002'\'',' 05 11 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-000002'\'',' + job_number=05 + test_step=11 + escape_command ''\''--name'\'', '\''jaeger-span-000002'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-000002'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-000002'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-000002'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=05 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./11-check-indices.yaml + JOB_NUMBER=05 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./11-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' 06 12 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + job_number=06 + test_step=12 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=06 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./12-check-indices.yaml + JOB_NUMBER=06 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./12-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.1"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.1"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.1"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.1 ++ version_ge 5.8.1 5.4 +++ echo 5.8.1 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.1 == 5.8.1 + '[' -n '' ']' + skip_test es-rollover-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-rollover-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-rollover-autoprov + '[' es-rollover-autoprov '!=' _build ']' + cd .. + rm -rf es-rollover-managed + warning 'es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + skip_test es-spark-dependencies 'This test is not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=es-spark-dependencies + message='This test is not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + rm -rf es-spark-dependencies + warning 'es-spark-dependencies: This test is not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-spark-dependencies: This test is not supported in OpenShift\e[0m' WAR: es-spark-dependencies: This test is not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running elasticsearch E2E tests' Running elasticsearch E2E tests + cd tests/e2e/elasticsearch/_build + set +e + KUBECONFIG=/tmp/kubeconfig-1441553849 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/es-from-aio-to-production === PAUSE kuttl/harness/es-from-aio-to-production === RUN kuttl/harness/es-increasing-replicas === PAUSE kuttl/harness/es-increasing-replicas === RUN kuttl/harness/es-index-cleaner-autoprov === PAUSE kuttl/harness/es-index-cleaner-autoprov === RUN kuttl/harness/es-multiinstance === PAUSE kuttl/harness/es-multiinstance === RUN kuttl/harness/es-rollover-autoprov === PAUSE kuttl/harness/es-rollover-autoprov === RUN kuttl/harness/es-simple-prod === PAUSE kuttl/harness/es-simple-prod === CONT kuttl/harness/artifacts logger.go:42: 07:33:25 | artifacts | Creating namespace: kuttl-test-wired-tick logger.go:42: 07:33:26 | artifacts | artifacts events from ns kuttl-test-wired-tick: logger.go:42: 07:33:26 | artifacts | Deleting namespace: kuttl-test-wired-tick === CONT kuttl/harness/es-multiinstance logger.go:42: 07:33:32 | es-multiinstance | Ignoring 03-create-second-instance.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:33:32 | es-multiinstance | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:33:32 | es-multiinstance | Creating namespace: kuttl-test-tops-grubworm logger.go:42: 07:33:32 | es-multiinstance/0-clear-namespace | starting test step 0-clear-namespace logger.go:42: 07:33:32 | es-multiinstance/0-clear-namespace | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --ignore-not-found=true] logger.go:42: 07:33:32 | es-multiinstance/0-clear-namespace | test step completed 0-clear-namespace logger.go:42: 07:33:32 | es-multiinstance/1-install | starting test step 1-install logger.go:42: 07:33:32 | es-multiinstance/1-install | Jaeger:kuttl-test-tops-grubworm/instance-1 created logger.go:42: 07:34:21 | es-multiinstance/1-install | test step completed 1-install logger.go:42: 07:34:21 | es-multiinstance/2-create-namespace | starting test step 2-create-namespace logger.go:42: 07:34:21 | es-multiinstance/2-create-namespace | running command: [sh -c kubectl create namespace jaeger-e2e-multiinstance-test] logger.go:42: 07:34:21 | es-multiinstance/2-create-namespace | namespace/jaeger-e2e-multiinstance-test created logger.go:42: 07:34:21 | es-multiinstance/2-create-namespace | test step completed 2-create-namespace logger.go:42: 07:34:21 | es-multiinstance/3-create-second-instance | starting test step 3-create-second-instance logger.go:42: 07:34:21 | es-multiinstance/3-create-second-instance | running command: [sh -c kubectl apply -f ./01-install.yaml -n jaeger-e2e-multiinstance-test] logger.go:42: 07:34:23 | es-multiinstance/3-create-second-instance | jaeger.jaegertracing.io/instance-1 created logger.go:42: 07:34:23 | es-multiinstance/3-create-second-instance | running command: [sh -c /tmp/jaeger-tests/bin/kubectl-kuttl assert ./01-assert.yaml -n jaeger-e2e-multiinstance-test --timeout 1000] logger.go:42: 07:35:15 | es-multiinstance/3-create-second-instance | assert is valid logger.go:42: 07:35:15 | es-multiinstance/3-create-second-instance | test step completed 3-create-second-instance logger.go:42: 07:35:15 | es-multiinstance/4-check-secrets | starting test step 4-check-secrets logger.go:42: 07:35:15 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n $NAMESPACE > secret1] logger.go:42: 07:35:15 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n jaeger-e2e-multiinstance-test > secret2] logger.go:42: 07:35:15 | es-multiinstance/4-check-secrets | running command: [sh -c cmp --silent secret1 secret2 || exit 0] logger.go:42: 07:35:15 | es-multiinstance/4-check-secrets | test step completed 4-check-secrets logger.go:42: 07:35:15 | es-multiinstance/5-delete | starting test step 5-delete logger.go:42: 07:35:15 | es-multiinstance/5-delete | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --wait=false] logger.go:42: 07:35:15 | es-multiinstance/5-delete | namespace "jaeger-e2e-multiinstance-test" deleted logger.go:42: 07:35:15 | es-multiinstance/5-delete | test step completed 5-delete logger.go:42: 07:35:15 | es-multiinstance | es-multiinstance events from ns kuttl-test-tops-grubworm: logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:39 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltesttopsgrubworminstance1-1-56f57f8bd7 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltesttopsgrubworminstance1-1-56f57f8bs65j replicaset-controller logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopsgrubworminstance1-1-56f57f8bs65j Binding Scheduled Successfully assigned kuttl-test-tops-grubworm/elasticsearch-cdm-kuttltesttopsgrubworminstance1-1-56f57f8bs65j to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:39 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltesttopsgrubworminstance1-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltesttopsgrubworminstance1-1-56f57f8bd7 to 1 deployment-controller logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:39 +0000 UTC Normal Pod instance-1-collector-587c8c8d45-ltzc8 Binding Scheduled Successfully assigned kuttl-test-tops-grubworm/instance-1-collector-587c8c8d45-ltzc8 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:39 +0000 UTC Normal ReplicaSet.apps instance-1-collector-587c8c8d45 SuccessfulCreate Created pod: instance-1-collector-587c8c8d45-ltzc8 replicaset-controller logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:39 +0000 UTC Normal Deployment.apps instance-1-collector ScalingReplicaSet Scaled up replica set instance-1-collector-587c8c8d45 to 1 deployment-controller logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopsgrubworminstance1-1-56f57f8bs65j AddedInterface Add eth0 [10.129.2.19/23] from ovn-kubernetes logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopsgrubworminstance1-1-56f57f8bs65j.spec.containers{elasticsearch} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:40 +0000 UTC Normal Pod instance-1-collector-587c8c8d45-ltzc8 AddedInterface Add eth0 [10.128.2.19/23] from ovn-kubernetes logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:40 +0000 UTC Normal Pod instance-1-collector-587c8c8d45-ltzc8.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:40 +0000 UTC Normal Pod instance-1-query-857fd78bfc-v4gcb Binding Scheduled Successfully assigned kuttl-test-tops-grubworm/instance-1-query-857fd78bfc-v4gcb to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:40 +0000 UTC Normal Pod instance-1-query-857fd78bfc-v4gcb AddedInterface Add eth0 [10.131.0.19/23] from ovn-kubernetes logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:40 +0000 UTC Normal Pod instance-1-query-857fd78bfc-v4gcb.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:40 +0000 UTC Normal ReplicaSet.apps instance-1-query-857fd78bfc SuccessfulCreate Created pod: instance-1-query-857fd78bfc-v4gcb replicaset-controller logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:40 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-857fd78bfc to 1 deployment-controller logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopsgrubworminstance1-1-56f57f8bs65j.spec.containers{elasticsearch} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" in 4.951s (4.951s including waiting) kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopsgrubworminstance1-1-56f57f8bs65j.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopsgrubworminstance1-1-56f57f8bs65j.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopsgrubworminstance1-1-56f57f8bs65j.spec.containers{proxy} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:45 +0000 UTC Normal Pod instance-1-collector-587c8c8d45-ltzc8.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" in 5.054s (5.054s including waiting) kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:45 +0000 UTC Normal Pod instance-1-collector-587c8c8d45-ltzc8.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:45 +0000 UTC Normal Pod instance-1-collector-587c8c8d45-ltzc8.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:46 +0000 UTC Warning Pod instance-1-collector-587c8c8d45-ltzc8.spec.containers{jaeger-collector} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:46 +0000 UTC Normal Pod instance-1-query-857fd78bfc-v4gcb.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" in 5.203s (5.203s including waiting) kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:46 +0000 UTC Normal Pod instance-1-query-857fd78bfc-v4gcb.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:46 +0000 UTC Normal Pod instance-1-query-857fd78bfc-v4gcb.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:46 +0000 UTC Normal Pod instance-1-query-857fd78bfc-v4gcb.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopsgrubworminstance1-1-56f57f8bs65j.spec.containers{proxy} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" in 2.729s (2.729s including waiting) kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopsgrubworminstance1-1-56f57f8bs65j.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopsgrubworminstance1-1-56f57f8bs65j.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:50 +0000 UTC Normal Pod instance-1-collector-587c8c8d45-ltzc8.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:50 +0000 UTC Normal Pod instance-1-query-857fd78bfc-v4gcb.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" in 4.17s (4.171s including waiting) kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:50 +0000 UTC Normal Pod instance-1-query-857fd78bfc-v4gcb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:50 +0000 UTC Normal Pod instance-1-query-857fd78bfc-v4gcb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:50 +0000 UTC Normal Pod instance-1-query-857fd78bfc-v4gcb.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:54 +0000 UTC Normal Pod instance-1-query-857fd78bfc-v4gcb.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" in 3.585s (3.585s including waiting) kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:54 +0000 UTC Normal Pod instance-1-query-857fd78bfc-v4gcb.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:54 +0000 UTC Normal Pod instance-1-query-857fd78bfc-v4gcb.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:55 +0000 UTC Warning Pod instance-1-collector-587c8c8d45-ltzc8.spec.containers{jaeger-collector} Unhealthy Liveness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:55 +0000 UTC Normal Pod instance-1-query-857fd78bfc-v4gcb.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:56 +0000 UTC Warning Pod instance-1-collector-587c8c8d45-ltzc8.spec.containers{jaeger-collector} BackOff Back-off restarting failed container jaeger-collector in pod instance-1-collector-587c8c8d45-ltzc8_kuttl-test-tops-grubworm(4ea0bde5-bfd7-4f9b-b006-ededf644ac4d) kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:33:56 +0000 UTC Warning Pod instance-1-query-857fd78bfc-v4gcb.spec.containers{jaeger-query} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:34:00 +0000 UTC Warning Pod elasticsearch-cdm-kuttltesttopsgrubworminstance1-1-56f57f8bs65j.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:34:00 +0000 UTC Warning Pod instance-1-query-857fd78bfc-v4gcb.spec.containers{jaeger-query} Unhealthy Readiness probe failed: Get "http://10.131.0.19:16687/": dial tcp 10.131.0.19:16687: connect: connection refused kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:34:01 +0000 UTC Warning Pod instance-1-query-857fd78bfc-v4gcb.spec.containers{jaeger-query} BackOff Back-off restarting failed container jaeger-query in pod instance-1-query-857fd78bfc-v4gcb_kuttl-test-tops-grubworm(474feabd-a28f-4dbe-9a8f-a7051260d73c) kubelet logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:34:37 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:34:37 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:34:37 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:34:52 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:34:52 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod instance-1-collector-587c8c8d45-ltzc8 horizontal-pod-autoscaler logger.go:42: 07:35:15 | es-multiinstance | 2023-12-25 07:34:52 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:35:15 | es-multiinstance | Deleting namespace: kuttl-test-tops-grubworm === CONT kuttl/harness/es-simple-prod logger.go:42: 07:35:23 | es-simple-prod | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:35:23 | es-simple-prod | Creating namespace: kuttl-test-prepared-marlin logger.go:42: 07:35:23 | es-simple-prod | es-simple-prod events from ns kuttl-test-prepared-marlin: logger.go:42: 07:35:23 | es-simple-prod | Deleting namespace: kuttl-test-prepared-marlin === CONT kuttl/harness/es-rollover-autoprov logger.go:42: 07:35:29 | es-rollover-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:35:29 | es-rollover-autoprov | Creating namespace: kuttl-test-select-insect logger.go:42: 07:35:29 | es-rollover-autoprov/1-install | starting test step 1-install logger.go:42: 07:35:29 | es-rollover-autoprov/1-install | Jaeger:kuttl-test-select-insect/my-jaeger created logger.go:42: 07:36:05 | es-rollover-autoprov/1-install | test step completed 1-install logger.go:42: 07:36:05 | es-rollover-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:36:05 | es-rollover-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:36:08 | es-rollover-autoprov/2-report-spans | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:36:15 | es-rollover-autoprov/2-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:36:15 | es-rollover-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:36:16 | es-rollover-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:36:42 | es-rollover-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 07:36:42 | es-rollover-autoprov/3-check-indices | starting test step 3-check-indices logger.go:42: 07:36:42 | es-rollover-autoprov/3-check-indices | Job:kuttl-test-select-insect/00-check-indices created logger.go:42: 07:36:46 | es-rollover-autoprov/3-check-indices | test step completed 3-check-indices logger.go:42: 07:36:46 | es-rollover-autoprov/4-check-indices | starting test step 4-check-indices logger.go:42: 07:36:46 | es-rollover-autoprov/4-check-indices | Job:kuttl-test-select-insect/01-check-indices created logger.go:42: 07:36:51 | es-rollover-autoprov/4-check-indices | test step completed 4-check-indices logger.go:42: 07:36:51 | es-rollover-autoprov/5-install | starting test step 5-install logger.go:42: 07:36:51 | es-rollover-autoprov/5-install | Jaeger:kuttl-test-select-insect/my-jaeger updated logger.go:42: 07:37:04 | es-rollover-autoprov/5-install | test step completed 5-install logger.go:42: 07:37:04 | es-rollover-autoprov/6-report-spans | starting test step 6-report-spans logger.go:42: 07:37:04 | es-rollover-autoprov/6-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:37:11 | es-rollover-autoprov/6-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 JOB_NUMBER=02 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-02-job.yaml] logger.go:42: 07:37:12 | es-rollover-autoprov/6-report-spans | running command: [sh -c kubectl apply -f report-span-02-job.yaml -n $NAMESPACE] logger.go:42: 07:37:12 | es-rollover-autoprov/6-report-spans | job.batch/02-report-span created logger.go:42: 07:37:37 | es-rollover-autoprov/6-report-spans | test step completed 6-report-spans logger.go:42: 07:37:37 | es-rollover-autoprov/7-check-indices | starting test step 7-check-indices logger.go:42: 07:37:37 | es-rollover-autoprov/7-check-indices | Job:kuttl-test-select-insect/02-check-indices created logger.go:42: 07:37:41 | es-rollover-autoprov/7-check-indices | test step completed 7-check-indices logger.go:42: 07:37:41 | es-rollover-autoprov/8-check-indices | starting test step 8-check-indices logger.go:42: 07:37:41 | es-rollover-autoprov/8-check-indices | Job:kuttl-test-select-insect/03-check-indices created logger.go:42: 07:37:44 | es-rollover-autoprov/8-check-indices | test step completed 8-check-indices logger.go:42: 07:37:44 | es-rollover-autoprov/9-check-indices | starting test step 9-check-indices logger.go:42: 07:37:44 | es-rollover-autoprov/9-check-indices | Job:kuttl-test-select-insect/04-check-indices created logger.go:42: 07:37:47 | es-rollover-autoprov/9-check-indices | test step completed 9-check-indices logger.go:42: 07:37:47 | es-rollover-autoprov/10-report-spans | starting test step 10-report-spans logger.go:42: 07:37:47 | es-rollover-autoprov/10-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:37:56 | es-rollover-autoprov/10-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 JOB_NUMBER=03 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-03-job.yaml] logger.go:42: 07:37:57 | es-rollover-autoprov/10-report-spans | running command: [sh -c kubectl apply -f report-span-03-job.yaml -n $NAMESPACE] logger.go:42: 07:37:57 | es-rollover-autoprov/10-report-spans | job.batch/03-report-span created logger.go:42: 07:38:20 | es-rollover-autoprov/10-report-spans | test step completed 10-report-spans logger.go:42: 07:38:20 | es-rollover-autoprov/11-check-indices | starting test step 11-check-indices logger.go:42: 07:38:20 | es-rollover-autoprov/11-check-indices | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob my-jaeger-es-rollover --namespace $NAMESPACE] logger.go:42: 07:38:31 | es-rollover-autoprov/11-check-indices | time="2023-12-25T07:38:31Z" level=debug msg="Checking if the my-jaeger-es-rollover CronJob exists" logger.go:42: 07:38:31 | es-rollover-autoprov/11-check-indices | time="2023-12-25T07:38:31Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 07:38:31 | es-rollover-autoprov/11-check-indices | time="2023-12-25T07:38:31Z" level=info msg="Cronjob my-jaeger-es-rollover found successfully" logger.go:42: 07:38:31 | es-rollover-autoprov/11-check-indices | time="2023-12-25T07:38:31Z" level=debug msg="Waiting for the next scheduled job from my-jaeger-es-rollover cronjob" logger.go:42: 07:38:31 | es-rollover-autoprov/11-check-indices | time="2023-12-25T07:38:31Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:38:41 | es-rollover-autoprov/11-check-indices | time="2023-12-25T07:38:41Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:38:51 | es-rollover-autoprov/11-check-indices | time="2023-12-25T07:38:51Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:39:01 | es-rollover-autoprov/11-check-indices | time="2023-12-25T07:39:01Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:39:11 | es-rollover-autoprov/11-check-indices | time="2023-12-25T07:39:11Z" level=info msg="Job of owner my-jaeger-es-rollover succeeded after my-jaeger-es-rollover 40.049294401s" logger.go:42: 07:39:11 | es-rollover-autoprov/11-check-indices | Job:kuttl-test-select-insect/05-check-indices created logger.go:42: 07:39:15 | es-rollover-autoprov/11-check-indices | test step completed 11-check-indices logger.go:42: 07:39:15 | es-rollover-autoprov/12-check-indices | starting test step 12-check-indices logger.go:42: 07:39:15 | es-rollover-autoprov/12-check-indices | Job:kuttl-test-select-insect/06-check-indices created logger.go:42: 07:39:19 | es-rollover-autoprov/12-check-indices | test step completed 12-check-indices logger.go:42: 07:39:19 | es-rollover-autoprov | es-rollover-autoprov events from ns kuttl-test-select-insect: logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:35:35 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestselectinsectmyjaeger-1-5449bc5d49 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestselectinsectmyjaeger-1-5449bc5djmcfk replicaset-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:35:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestselectinsectmyjaeger-1-5449bc5djmcfk Binding Scheduled Successfully assigned kuttl-test-select-insect/elasticsearch-cdm-kuttltestselectinsectmyjaeger-1-5449bc5djmcfk to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:35:35 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestselectinsectmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestselectinsectmyjaeger-1-5449bc5d49 to 1 deployment-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:35:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestselectinsectmyjaeger-1-5449bc5djmcfk AddedInterface Add eth0 [10.129.2.20/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:35:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestselectinsectmyjaeger-1-5449bc5djmcfk.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:35:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestselectinsectmyjaeger-1-5449bc5djmcfk.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:35:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestselectinsectmyjaeger-1-5449bc5djmcfk.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:35:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestselectinsectmyjaeger-1-5449bc5djmcfk.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:35:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestselectinsectmyjaeger-1-5449bc5djmcfk.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:35:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestselectinsectmyjaeger-1-5449bc5djmcfk.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:35:46 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestselectinsectmyjaeger-1-5449bc5djmcfk.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:35:51 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestselectinsectmyjaeger-1-5449bc5djmcfk.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:02 +0000 UTC Normal Pod my-jaeger-collector-c996ddf6d-4bfcb Binding Scheduled Successfully assigned kuttl-test-select-insect/my-jaeger-collector-c996ddf6d-4bfcb to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:02 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-c996ddf6d SuccessfulCreate Created pod: my-jaeger-collector-c996ddf6d-4bfcb replicaset-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:02 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-c996ddf6d to 1 deployment-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:02 +0000 UTC Normal Pod my-jaeger-query-66cd4c749d-w2zzt Binding Scheduled Successfully assigned kuttl-test-select-insect/my-jaeger-query-66cd4c749d-w2zzt to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:02 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-66cd4c749d SuccessfulCreate Created pod: my-jaeger-query-66cd4c749d-w2zzt replicaset-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:02 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-66cd4c749d to 1 deployment-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:03 +0000 UTC Normal Pod my-jaeger-collector-c996ddf6d-4bfcb AddedInterface Add eth0 [10.128.2.21/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:03 +0000 UTC Normal Pod my-jaeger-collector-c996ddf6d-4bfcb.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:03 +0000 UTC Normal Pod my-jaeger-collector-c996ddf6d-4bfcb.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:03 +0000 UTC Normal Pod my-jaeger-collector-c996ddf6d-4bfcb.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:03 +0000 UTC Normal Pod my-jaeger-query-66cd4c749d-w2zzt AddedInterface Add eth0 [10.131.0.22/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:03 +0000 UTC Normal Pod my-jaeger-query-66cd4c749d-w2zzt.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:03 +0000 UTC Normal Pod my-jaeger-query-66cd4c749d-w2zzt.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:03 +0000 UTC Normal Pod my-jaeger-query-66cd4c749d-w2zzt.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:03 +0000 UTC Normal Pod my-jaeger-query-66cd4c749d-w2zzt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:03 +0000 UTC Normal Pod my-jaeger-query-66cd4c749d-w2zzt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:03 +0000 UTC Normal Pod my-jaeger-query-66cd4c749d-w2zzt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:03 +0000 UTC Normal Pod my-jaeger-query-66cd4c749d-w2zzt.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:03 +0000 UTC Normal Pod my-jaeger-query-66cd4c749d-w2zzt.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:03 +0000 UTC Normal Pod my-jaeger-query-66cd4c749d-w2zzt.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:10 +0000 UTC Normal Pod my-jaeger-query-66cd4c749d-w2zzt.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:10 +0000 UTC Normal Pod my-jaeger-query-66cd4c749d-w2zzt.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:10 +0000 UTC Normal Pod my-jaeger-query-66cd4c749d-w2zzt.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:10 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-66cd4c749d SuccessfulDelete Deleted pod: my-jaeger-query-66cd4c749d-w2zzt replicaset-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:10 +0000 UTC Normal Pod my-jaeger-query-84d6847fd5-72cbz Binding Scheduled Successfully assigned kuttl-test-select-insect/my-jaeger-query-84d6847fd5-72cbz to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:10 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-84d6847fd5 SuccessfulCreate Created pod: my-jaeger-query-84d6847fd5-72cbz replicaset-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:10 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-66cd4c749d to 0 from 1 deployment-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:10 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-84d6847fd5 to 1 deployment-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:11 +0000 UTC Normal Pod my-jaeger-query-84d6847fd5-72cbz AddedInterface Add eth0 [10.131.0.23/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:11 +0000 UTC Normal Pod my-jaeger-query-84d6847fd5-72cbz.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:11 +0000 UTC Normal Pod my-jaeger-query-84d6847fd5-72cbz.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:11 +0000 UTC Normal Pod my-jaeger-query-84d6847fd5-72cbz.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:11 +0000 UTC Normal Pod my-jaeger-query-84d6847fd5-72cbz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:11 +0000 UTC Normal Pod my-jaeger-query-84d6847fd5-72cbz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:11 +0000 UTC Normal Pod my-jaeger-query-84d6847fd5-72cbz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:11 +0000 UTC Normal Pod my-jaeger-query-84d6847fd5-72cbz.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:11 +0000 UTC Normal Pod my-jaeger-query-84d6847fd5-72cbz.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:11 +0000 UTC Normal Pod my-jaeger-query-84d6847fd5-72cbz.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:16 +0000 UTC Normal Pod 00-report-span-nznfr Binding Scheduled Successfully assigned kuttl-test-select-insect/00-report-span-nznfr to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:16 +0000 UTC Normal Pod 00-report-span-nznfr AddedInterface Add eth0 [10.128.2.22/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:16 +0000 UTC Normal Pod 00-report-span-nznfr.spec.containers{asserts-container} Pulling Pulling image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:16 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-nznfr job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:17 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:17 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:17 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:19 +0000 UTC Normal Pod 00-report-span-nznfr.spec.containers{asserts-container} Pulled Successfully pulled image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" in 2.525s (2.525s including waiting) kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:19 +0000 UTC Normal Pod 00-report-span-nznfr.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:19 +0000 UTC Normal Pod 00-report-span-nznfr.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:42 +0000 UTC Normal Pod 00-check-indices-cmg74 Binding Scheduled Successfully assigned kuttl-test-select-insect/00-check-indices-cmg74 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:42 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-cmg74 job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:42 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:43 +0000 UTC Normal Pod 00-check-indices-cmg74 AddedInterface Add eth0 [10.128.2.23/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:43 +0000 UTC Normal Pod 00-check-indices-cmg74.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:43 +0000 UTC Normal Pod 00-check-indices-cmg74.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:43 +0000 UTC Normal Pod 00-check-indices-cmg74.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:46 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:47 +0000 UTC Normal Pod 01-check-indices-n5vwg Binding Scheduled Successfully assigned kuttl-test-select-insect/01-check-indices-n5vwg to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:47 +0000 UTC Normal Pod 01-check-indices-n5vwg AddedInterface Add eth0 [10.128.2.24/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:47 +0000 UTC Normal Pod 01-check-indices-n5vwg.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:47 +0000 UTC Normal Pod 01-check-indices-n5vwg.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:47 +0000 UTC Normal Pod 01-check-indices-n5vwg.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:47 +0000 UTC Normal Job.batch 01-check-indices SuccessfulCreate Created pod: 01-check-indices-n5vwg job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:50 +0000 UTC Normal Job.batch 01-check-indices Completed Job completed job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:52 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-zs7xj Binding Scheduled Successfully assigned kuttl-test-select-insect/my-jaeger-es-rollover-create-mapping-zs7xj to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:52 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-zs7xj AddedInterface Add eth0 [10.128.2.25/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:52 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-zs7xj.spec.containers{my-jaeger-es-rollover-create-mapping} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:52 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping SuccessfulCreate Created pod: my-jaeger-es-rollover-create-mapping-zs7xj job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:57 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-zs7xj.spec.containers{my-jaeger-es-rollover-create-mapping} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" in 4.495s (4.495s including waiting) kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:57 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-zs7xj.spec.containers{my-jaeger-es-rollover-create-mapping} Created Created container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:36:57 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-zs7xj.spec.containers{my-jaeger-es-rollover-create-mapping} Started Started container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:00 +0000 UTC Normal Pod my-jaeger-collector-c996ddf6d-4bfcb.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:00 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-c996ddf6d SuccessfulDelete Deleted pod: my-jaeger-collector-c996ddf6d-4bfcb replicaset-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:00 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-c996ddf6d to 0 from 1 deployment-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping Completed Job completed job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:00 +0000 UTC Normal Pod my-jaeger-query-84d6847fd5-72cbz.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:00 +0000 UTC Normal Pod my-jaeger-query-84d6847fd5-72cbz.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:00 +0000 UTC Normal Pod my-jaeger-query-84d6847fd5-72cbz.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:00 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-84d6847fd5 SuccessfulDelete Deleted pod: my-jaeger-query-84d6847fd5-72cbz replicaset-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:00 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-84d6847fd5 to 0 from 1 deployment-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:01 +0000 UTC Normal Pod my-jaeger-collector-65b8955d95-n5qhb Binding Scheduled Successfully assigned kuttl-test-select-insect/my-jaeger-collector-65b8955d95-n5qhb to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:01 +0000 UTC Normal Pod my-jaeger-collector-65b8955d95-n5qhb AddedInterface Add eth0 [10.128.2.26/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:01 +0000 UTC Normal Pod my-jaeger-collector-65b8955d95-n5qhb.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:01 +0000 UTC Normal Pod my-jaeger-collector-65b8955d95-n5qhb.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:01 +0000 UTC Normal Pod my-jaeger-collector-65b8955d95-n5qhb.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:01 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-65b8955d95 SuccessfulCreate Created pod: my-jaeger-collector-65b8955d95-n5qhb replicaset-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:01 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-65b8955d95 to 1 deployment-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:01 +0000 UTC Normal Pod my-jaeger-query-7cd8f7c484-wfjcb Binding Scheduled Successfully assigned kuttl-test-select-insect/my-jaeger-query-7cd8f7c484-wfjcb to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:01 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7cd8f7c484 SuccessfulCreate Created pod: my-jaeger-query-7cd8f7c484-wfjcb replicaset-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:01 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7cd8f7c484 to 1 deployment-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:02 +0000 UTC Normal Pod my-jaeger-query-7cd8f7c484-wfjcb AddedInterface Add eth0 [10.131.0.24/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:02 +0000 UTC Normal Pod my-jaeger-query-7cd8f7c484-wfjcb.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:02 +0000 UTC Normal Pod my-jaeger-query-7cd8f7c484-wfjcb.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:02 +0000 UTC Normal Pod my-jaeger-query-7cd8f7c484-wfjcb.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:02 +0000 UTC Normal Pod my-jaeger-query-7cd8f7c484-wfjcb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:02 +0000 UTC Normal Pod my-jaeger-query-7cd8f7c484-wfjcb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:02 +0000 UTC Normal Pod my-jaeger-query-7cd8f7c484-wfjcb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:02 +0000 UTC Normal Pod my-jaeger-query-7cd8f7c484-wfjcb.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:02 +0000 UTC Normal Pod my-jaeger-query-7cd8f7c484-wfjcb.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:02 +0000 UTC Normal Pod my-jaeger-query-7cd8f7c484-wfjcb.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:12 +0000 UTC Normal Pod 02-report-span-tw7ws Binding Scheduled Successfully assigned kuttl-test-select-insect/02-report-span-tw7ws to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:12 +0000 UTC Normal Job.batch 02-report-span SuccessfulCreate Created pod: 02-report-span-tw7ws job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:13 +0000 UTC Normal Pod 02-report-span-tw7ws AddedInterface Add eth0 [10.128.2.27/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:13 +0000 UTC Normal Pod 02-report-span-tw7ws.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:13 +0000 UTC Normal Pod 02-report-span-tw7ws.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:13 +0000 UTC Normal Pod 02-report-span-tw7ws.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:36 +0000 UTC Normal Job.batch 02-report-span Completed Job completed job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:37 +0000 UTC Normal Pod 02-check-indices-xqpm6 Binding Scheduled Successfully assigned kuttl-test-select-insect/02-check-indices-xqpm6 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:37 +0000 UTC Normal Pod 02-check-indices-xqpm6 AddedInterface Add eth0 [10.128.2.28/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:37 +0000 UTC Normal Pod 02-check-indices-xqpm6.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:37 +0000 UTC Normal Pod 02-check-indices-xqpm6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:37 +0000 UTC Normal Pod 02-check-indices-xqpm6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:37 +0000 UTC Normal Job.batch 02-check-indices SuccessfulCreate Created pod: 02-check-indices-xqpm6 job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:40 +0000 UTC Normal Job.batch 02-check-indices Completed Job completed job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:41 +0000 UTC Normal Pod 03-check-indices-pz28w Binding Scheduled Successfully assigned kuttl-test-select-insect/03-check-indices-pz28w to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:41 +0000 UTC Normal Job.batch 03-check-indices SuccessfulCreate Created pod: 03-check-indices-pz28w job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:42 +0000 UTC Normal Pod 03-check-indices-pz28w AddedInterface Add eth0 [10.128.2.29/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:42 +0000 UTC Normal Pod 03-check-indices-pz28w.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:42 +0000 UTC Normal Pod 03-check-indices-pz28w.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:42 +0000 UTC Normal Pod 03-check-indices-pz28w.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:44 +0000 UTC Normal Job.batch 03-check-indices Completed Job completed job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:44 +0000 UTC Normal Pod 04-check-indices-hlzj6 Binding Scheduled Successfully assigned kuttl-test-select-insect/04-check-indices-hlzj6 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:44 +0000 UTC Normal Pod 04-check-indices-hlzj6 AddedInterface Add eth0 [10.128.2.30/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:44 +0000 UTC Normal Pod 04-check-indices-hlzj6.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:44 +0000 UTC Normal Job.batch 04-check-indices SuccessfulCreate Created pod: 04-check-indices-hlzj6 job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:45 +0000 UTC Normal Pod 04-check-indices-hlzj6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:45 +0000 UTC Normal Pod 04-check-indices-hlzj6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:47 +0000 UTC Normal Job.batch 04-check-indices Completed Job completed job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:57 +0000 UTC Normal Pod 03-report-span-z2tnt Binding Scheduled Successfully assigned kuttl-test-select-insect/03-report-span-z2tnt to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:57 +0000 UTC Normal Pod 03-report-span-z2tnt AddedInterface Add eth0 [10.128.2.31/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:57 +0000 UTC Normal Pod 03-report-span-z2tnt.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:57 +0000 UTC Normal Pod 03-report-span-z2tnt.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:57 +0000 UTC Normal Pod 03-report-span-z2tnt.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:37:57 +0000 UTC Normal Job.batch 03-report-span SuccessfulCreate Created pod: 03-report-span-z2tnt job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28391498-hstn9 Binding Scheduled Successfully assigned kuttl-test-select-insect/my-jaeger-es-lookback-28391498-hstn9 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28391498-hstn9 AddedInterface Add eth0 [10.128.2.33/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28391498-hstn9.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28391498-hstn9.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28391498-hstn9.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28391498 SuccessfulCreate Created pod: my-jaeger-es-lookback-28391498-hstn9 job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28391498 cronjob-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28391498-h4hkv Binding Scheduled Successfully assigned kuttl-test-select-insect/my-jaeger-es-rollover-28391498-h4hkv to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28391498-h4hkv AddedInterface Add eth0 [10.128.2.32/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28391498-h4hkv.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28391498-h4hkv.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28391498-h4hkv.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28391498 SuccessfulCreate Created pod: my-jaeger-es-rollover-28391498-h4hkv job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28391498 cronjob-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:03 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:03 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-65b8955d95-n5qhb horizontal-pod-autoscaler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:03 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28391498 Completed Job completed job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28391498, status: Complete cronjob-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28391498 Completed Job completed job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28391498, status: Complete cronjob-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:38:20 +0000 UTC Normal Job.batch 03-report-span Completed Job completed job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28391499-z9r6f Binding Scheduled Successfully assigned kuttl-test-select-insect/my-jaeger-es-lookback-28391499-z9r6f to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28391499-z9r6f AddedInterface Add eth0 [10.128.2.35/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28391499-z9r6f.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28391499-z9r6f.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28391499-z9r6f.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28391499 SuccessfulCreate Created pod: my-jaeger-es-lookback-28391499-z9r6f job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28391499 cronjob-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28391499-kcz5d Binding Scheduled Successfully assigned kuttl-test-select-insect/my-jaeger-es-rollover-28391499-kcz5d to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28391499-kcz5d AddedInterface Add eth0 [10.128.2.34/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28391499-kcz5d.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28391499-kcz5d.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28391499-kcz5d.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28391499 SuccessfulCreate Created pod: my-jaeger-es-rollover-28391499-kcz5d job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28391499 cronjob-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28391499 Completed Job completed job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28391499, status: Complete cronjob-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28391499 Completed Job completed job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28391499, status: Complete cronjob-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:11 +0000 UTC Normal Pod 05-check-indices-2wjrd Binding Scheduled Successfully assigned kuttl-test-select-insect/05-check-indices-2wjrd to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:11 +0000 UTC Normal Pod 05-check-indices-2wjrd AddedInterface Add eth0 [10.128.2.36/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:11 +0000 UTC Normal Pod 05-check-indices-2wjrd.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:11 +0000 UTC Normal Job.batch 05-check-indices SuccessfulCreate Created pod: 05-check-indices-2wjrd job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:12 +0000 UTC Normal Pod 05-check-indices-2wjrd.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:12 +0000 UTC Normal Pod 05-check-indices-2wjrd.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:14 +0000 UTC Normal Job.batch 05-check-indices Completed Job completed job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:15 +0000 UTC Normal Pod 06-check-indices-6md6c Binding Scheduled Successfully assigned kuttl-test-select-insect/06-check-indices-6md6c to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:15 +0000 UTC Normal Job.batch 06-check-indices SuccessfulCreate Created pod: 06-check-indices-6md6c job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:16 +0000 UTC Normal Pod 06-check-indices-6md6c AddedInterface Add eth0 [10.128.2.37/23] from ovn-kubernetes logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:16 +0000 UTC Normal Pod 06-check-indices-6md6c.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:16 +0000 UTC Normal Pod 06-check-indices-6md6c.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:16 +0000 UTC Normal Pod 06-check-indices-6md6c.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:39:19 | es-rollover-autoprov | 2023-12-25 07:39:18 +0000 UTC Normal Job.batch 06-check-indices Completed Job completed job-controller logger.go:42: 07:39:19 | es-rollover-autoprov | Deleting namespace: kuttl-test-select-insect === CONT kuttl/harness/es-increasing-replicas logger.go:42: 07:39:28 | es-increasing-replicas | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:39:28 | es-increasing-replicas | Ignoring check-es-nodes.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:39:28 | es-increasing-replicas | Ignoring openshift-check-es-nodes.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:39:28 | es-increasing-replicas | Creating namespace: kuttl-test-becoming-aardvark logger.go:42: 07:39:28 | es-increasing-replicas/1-install | starting test step 1-install logger.go:42: 07:39:28 | es-increasing-replicas/1-install | Jaeger:kuttl-test-becoming-aardvark/simple-prod created logger.go:42: 07:40:04 | es-increasing-replicas/1-install | test step completed 1-install logger.go:42: 07:40:04 | es-increasing-replicas/2-install | starting test step 2-install logger.go:42: 07:40:04 | es-increasing-replicas/2-install | Jaeger:kuttl-test-becoming-aardvark/simple-prod updated logger.go:42: 07:40:17 | es-increasing-replicas/2-install | test step completed 2-install logger.go:42: 07:40:17 | es-increasing-replicas/3-smoke-test | starting test step 3-smoke-test logger.go:42: 07:40:17 | es-increasing-replicas/3-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:40:19 | es-increasing-replicas/3-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:40:26 | es-increasing-replicas/3-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:40:26 | es-increasing-replicas/3-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:40:26 | es-increasing-replicas/3-smoke-test | job.batch/report-span created logger.go:42: 07:40:27 | es-increasing-replicas/3-smoke-test | job.batch/check-span created logger.go:42: 07:40:38 | es-increasing-replicas/3-smoke-test | test step completed 3-smoke-test logger.go:42: 07:40:38 | es-increasing-replicas/4-install | starting test step 4-install logger.go:42: 07:40:38 | es-increasing-replicas/4-install | Jaeger:kuttl-test-becoming-aardvark/simple-prod updated logger.go:42: 07:40:38 | es-increasing-replicas/4-install | test step completed 4-install logger.go:42: 07:40:38 | es-increasing-replicas/5-check-es-nodes | starting test step 5-check-es-nodes logger.go:42: 07:40:38 | es-increasing-replicas/5-check-es-nodes | running command: [sh -c ./check-es-nodes.sh $NAMESPACE] logger.go:42: 07:40:38 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:40:38 | es-increasing-replicas/5-check-es-nodes | false logger.go:42: 07:40:38 | es-increasing-replicas/5-check-es-nodes | Error: no matches found logger.go:42: 07:40:43 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:40:43 | es-increasing-replicas/5-check-es-nodes | true logger.go:42: 07:40:43 | es-increasing-replicas/5-check-es-nodes | test step completed 5-check-es-nodes logger.go:42: 07:40:43 | es-increasing-replicas | es-increasing-replicas events from ns kuttl-test-becoming-aardvark: logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:39:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-1-77cdjw7 Binding Scheduled Successfully assigned kuttl-test-becoming-aardvark/elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-1-77cdjw7 to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:39:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-1-77cdjw7 AddedInterface Add eth0 [10.129.2.21/23] from ovn-kubernetes logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:39:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-1-77cdjw7.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:39:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-1-77cdjw7.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:39:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-1-77cdjw7.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:39:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-1-77cdjw7.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:39:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-1-77cdjw7.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:39:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-1-77cdjw7.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:39:35 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-1-77d5b569b7 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-1-77cdjw7 replicaset-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:39:35 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-1-77d5b569b7 to 1 deployment-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:39:45 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-1-77cdjw7.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:39:50 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-1-77cdjw7.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal Pod simple-prod-collector-797b6fd85b-8vzd9 Binding Scheduled Successfully assigned kuttl-test-becoming-aardvark/simple-prod-collector-797b6fd85b-8vzd9 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal Pod simple-prod-collector-797b6fd85b-8vzd9 AddedInterface Add eth0 [10.128.2.38/23] from ovn-kubernetes logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal Pod simple-prod-collector-797b6fd85b-8vzd9.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal Pod simple-prod-collector-797b6fd85b-8vzd9.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal Pod simple-prod-collector-797b6fd85b-8vzd9.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-797b6fd85b SuccessfulCreate Created pod: simple-prod-collector-797b6fd85b-8vzd9 replicaset-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-797b6fd85b to 1 deployment-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal Pod simple-prod-query-8698d7b665-bfm78 Binding Scheduled Successfully assigned kuttl-test-becoming-aardvark/simple-prod-query-8698d7b665-bfm78 to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal Pod simple-prod-query-8698d7b665-bfm78 AddedInterface Add eth0 [10.131.0.25/23] from ovn-kubernetes logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal Pod simple-prod-query-8698d7b665-bfm78.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal Pod simple-prod-query-8698d7b665-bfm78.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal Pod simple-prod-query-8698d7b665-bfm78.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal Pod simple-prod-query-8698d7b665-bfm78.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal Pod simple-prod-query-8698d7b665-bfm78.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal Pod simple-prod-query-8698d7b665-bfm78.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal Pod simple-prod-query-8698d7b665-bfm78.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal Pod simple-prod-query-8698d7b665-bfm78.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal Pod simple-prod-query-8698d7b665-bfm78.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal ReplicaSet.apps simple-prod-query-8698d7b665 SuccessfulCreate Created pod: simple-prod-query-8698d7b665-bfm78 replicaset-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:02 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-8698d7b665 to 1 deployment-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:06 +0000 UTC Normal Pod simple-prod-collector-797b6fd85b-2fwq2 Binding Scheduled Successfully assigned kuttl-test-becoming-aardvark/simple-prod-collector-797b6fd85b-2fwq2 to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:06 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-797b6fd85b SuccessfulCreate Created pod: simple-prod-collector-797b6fd85b-2fwq2 replicaset-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:06 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-797b6fd85b to 2 from 1 deployment-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:06 +0000 UTC Normal Pod simple-prod-query-8698d7b665-k7pfl Binding Scheduled Successfully assigned kuttl-test-becoming-aardvark/simple-prod-query-8698d7b665-k7pfl to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:06 +0000 UTC Normal ReplicaSet.apps simple-prod-query-8698d7b665 SuccessfulCreate Created pod: simple-prod-query-8698d7b665-k7pfl replicaset-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:06 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-8698d7b665 to 2 from 1 deployment-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:07 +0000 UTC Normal Pod simple-prod-collector-797b6fd85b-2fwq2 AddedInterface Add eth0 [10.129.2.22/23] from ovn-kubernetes logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:07 +0000 UTC Normal Pod simple-prod-collector-797b6fd85b-2fwq2.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:07 +0000 UTC Normal Pod simple-prod-query-8698d7b665-k7pfl AddedInterface Add eth0 [10.129.2.23/23] from ovn-kubernetes logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:07 +0000 UTC Normal Pod simple-prod-query-8698d7b665-k7pfl.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:10 +0000 UTC Normal Pod simple-prod-collector-797b6fd85b-2fwq2.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" in 3.113s (3.113s including waiting) kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:10 +0000 UTC Normal Pod simple-prod-collector-797b6fd85b-2fwq2.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:10 +0000 UTC Normal Pod simple-prod-collector-797b6fd85b-2fwq2.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:10 +0000 UTC Normal Pod simple-prod-query-8698d7b665-k7pfl.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" in 3.073s (3.073s including waiting) kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:10 +0000 UTC Normal Pod simple-prod-query-8698d7b665-k7pfl.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:10 +0000 UTC Normal Pod simple-prod-query-8698d7b665-k7pfl.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:10 +0000 UTC Normal Pod simple-prod-query-8698d7b665-k7pfl.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:13 +0000 UTC Normal Pod simple-prod-query-8698d7b665-k7pfl.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" in 2.914s (2.914s including waiting) kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:13 +0000 UTC Normal Pod simple-prod-query-8698d7b665-k7pfl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:13 +0000 UTC Normal Pod simple-prod-query-8698d7b665-k7pfl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:13 +0000 UTC Normal Pod simple-prod-query-8698d7b665-k7pfl.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:16 +0000 UTC Normal Pod simple-prod-query-8698d7b665-k7pfl.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" in 2.61s (2.61s including waiting) kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:16 +0000 UTC Normal Pod simple-prod-query-8698d7b665-k7pfl.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:16 +0000 UTC Normal Pod simple-prod-query-8698d7b665-k7pfl.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:21 +0000 UTC Normal Pod simple-prod-query-8698d7b665-bfm78.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:21 +0000 UTC Normal Pod simple-prod-query-8698d7b665-bfm78.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:21 +0000 UTC Normal Pod simple-prod-query-8698d7b665-bfm78.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:21 +0000 UTC Normal Pod simple-prod-query-8698d7b665-k7pfl.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:21 +0000 UTC Normal Pod simple-prod-query-8698d7b665-k7pfl.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:21 +0000 UTC Normal Pod simple-prod-query-8698d7b665-k7pfl.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:21 +0000 UTC Normal ReplicaSet.apps simple-prod-query-8698d7b665 SuccessfulDelete Deleted pod: simple-prod-query-8698d7b665-k7pfl replicaset-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:21 +0000 UTC Normal ReplicaSet.apps simple-prod-query-8698d7b665 SuccessfulDelete Deleted pod: simple-prod-query-8698d7b665-bfm78 replicaset-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:21 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-8698d7b665 to 0 from 2 deployment-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:22 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-cdlfm Binding Scheduled Successfully assigned kuttl-test-becoming-aardvark/simple-prod-query-69dcb49d76-cdlfm to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:22 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-cdlfm AddedInterface Add eth0 [10.129.2.24/23] from ovn-kubernetes logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:22 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-cdlfm.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:22 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-mq86m Binding Scheduled Successfully assigned kuttl-test-becoming-aardvark/simple-prod-query-69dcb49d76-mq86m to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:22 +0000 UTC Normal ReplicaSet.apps simple-prod-query-69dcb49d76 SuccessfulCreate Created pod: simple-prod-query-69dcb49d76-mq86m replicaset-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:22 +0000 UTC Normal ReplicaSet.apps simple-prod-query-69dcb49d76 SuccessfulCreate Created pod: simple-prod-query-69dcb49d76-cdlfm replicaset-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:22 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-69dcb49d76 to 2 deployment-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:23 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-cdlfm.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:23 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-cdlfm.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:23 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-cdlfm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:23 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-cdlfm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:23 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-cdlfm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:23 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-cdlfm.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:23 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-cdlfm.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:23 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-cdlfm.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:23 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-mq86m AddedInterface Add eth0 [10.131.0.26/23] from ovn-kubernetes logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:23 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-mq86m.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:23 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-mq86m.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:23 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-mq86m.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:23 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-mq86m.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:23 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-mq86m.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:23 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-mq86m.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:23 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-mq86m.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:23 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-mq86m.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:23 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-mq86m.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:26 +0000 UTC Normal Pod report-span-vlfzv Binding Scheduled Successfully assigned kuttl-test-becoming-aardvark/report-span-vlfzv to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:26 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-vlfzv job-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:27 +0000 UTC Normal Pod check-span-k6tnf Binding Scheduled Successfully assigned kuttl-test-becoming-aardvark/check-span-k6tnf to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:27 +0000 UTC Normal Pod check-span-k6tnf AddedInterface Add eth0 [10.128.2.40/23] from ovn-kubernetes logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:27 +0000 UTC Normal Pod check-span-k6tnf.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:27 +0000 UTC Normal Pod check-span-k6tnf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:27 +0000 UTC Normal Pod check-span-k6tnf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:27 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-k6tnf job-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:27 +0000 UTC Normal Pod report-span-vlfzv AddedInterface Add eth0 [10.128.2.39/23] from ovn-kubernetes logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:27 +0000 UTC Normal Pod report-span-vlfzv.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:27 +0000 UTC Normal Pod report-span-vlfzv.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:27 +0000 UTC Normal Pod report-span-vlfzv.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:37 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-2-5b4ptdv Binding Scheduled Successfully assigned kuttl-test-becoming-aardvark/elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-2-5b4ptdv to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:39 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-2-5bbc7d676f SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-2-5b4ptdv replicaset-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:39 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-2 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-2-5bbc7d676f to 1 deployment-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:39 +0000 UTC Normal Pod simple-prod-collector-797b6fd85b-2fwq2.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:39 +0000 UTC Normal Pod simple-prod-collector-797b6fd85b-8vzd9.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:39 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-797b6fd85b SuccessfulDelete Deleted pod: simple-prod-collector-797b6fd85b-2fwq2 replicaset-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:39 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-797b6fd85b SuccessfulDelete Deleted pod: simple-prod-collector-797b6fd85b-8vzd9 replicaset-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:39 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled down replica set simple-prod-collector-797b6fd85b to 0 from 2 deployment-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:39 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-cdlfm.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:39 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-cdlfm.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:39 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-cdlfm.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:39 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-mq86m.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:39 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-mq86m.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:39 +0000 UTC Normal Pod simple-prod-query-69dcb49d76-mq86m.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:39 +0000 UTC Normal ReplicaSet.apps simple-prod-query-69dcb49d76 SuccessfulDelete Deleted pod: simple-prod-query-69dcb49d76-mq86m replicaset-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:39 +0000 UTC Normal ReplicaSet.apps simple-prod-query-69dcb49d76 SuccessfulDelete Deleted pod: simple-prod-query-69dcb49d76-cdlfm replicaset-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:39 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-69dcb49d76 to 0 from 2 deployment-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-2-5b4ptdv AddedInterface Add eth0 [10.128.2.41/23] from ovn-kubernetes logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-2-5b4ptdv.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-2-5b4ptdv.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-2-5b4ptdv.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-2-5b4ptdv.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-2-5b4ptdv.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingaardvarksimpleprod-2-5b4ptdv.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:40 +0000 UTC Normal Pod simple-prod-collector-c56fdcf6c-575v9 Binding Scheduled Successfully assigned kuttl-test-becoming-aardvark/simple-prod-collector-c56fdcf6c-575v9 to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:40 +0000 UTC Normal Pod simple-prod-collector-c56fdcf6c-b5h5f Binding Scheduled Successfully assigned kuttl-test-becoming-aardvark/simple-prod-collector-c56fdcf6c-b5h5f to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:40 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-c56fdcf6c SuccessfulCreate Created pod: simple-prod-collector-c56fdcf6c-b5h5f replicaset-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:40 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-c56fdcf6c SuccessfulCreate Created pod: simple-prod-collector-c56fdcf6c-575v9 replicaset-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:40 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-c56fdcf6c to 2 deployment-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:40 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-9bq48 Binding Scheduled Successfully assigned kuttl-test-becoming-aardvark/simple-prod-query-5c9d97c4fc-9bq48 to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:40 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-h2w5f Binding Scheduled Successfully assigned kuttl-test-becoming-aardvark/simple-prod-query-5c9d97c4fc-h2w5f to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:40 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5c9d97c4fc SuccessfulCreate Created pod: simple-prod-query-5c9d97c4fc-h2w5f replicaset-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:40 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5c9d97c4fc SuccessfulCreate Created pod: simple-prod-query-5c9d97c4fc-9bq48 replicaset-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:40 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-5c9d97c4fc to 2 deployment-controller logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-collector-c56fdcf6c-575v9 AddedInterface Add eth0 [10.129.2.26/23] from ovn-kubernetes logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-collector-c56fdcf6c-575v9.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-collector-c56fdcf6c-575v9.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-collector-c56fdcf6c-575v9.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-collector-c56fdcf6c-b5h5f AddedInterface Add eth0 [10.131.0.28/23] from ovn-kubernetes logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-collector-c56fdcf6c-b5h5f.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-collector-c56fdcf6c-b5h5f.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-collector-c56fdcf6c-b5h5f.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-9bq48 AddedInterface Add eth0 [10.129.2.25/23] from ovn-kubernetes logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-9bq48.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-9bq48.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-9bq48.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-9bq48.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-9bq48.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-9bq48.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-9bq48.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-9bq48.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-9bq48.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-h2w5f AddedInterface Add eth0 [10.131.0.27/23] from ovn-kubernetes logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-h2w5f.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-h2w5f.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-h2w5f.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-h2w5f.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-h2w5f.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-h2w5f.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-h2w5f.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-h2w5f.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:40:43 | es-increasing-replicas | 2023-12-25 07:40:41 +0000 UTC Normal Pod simple-prod-query-5c9d97c4fc-h2w5f.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:40:43 | es-increasing-replicas | Deleting namespace: kuttl-test-becoming-aardvark === CONT kuttl/harness/es-index-cleaner-autoprov logger.go:42: 07:41:16 | es-index-cleaner-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:41:16 | es-index-cleaner-autoprov | Creating namespace: kuttl-test-classic-ghost logger.go:42: 07:41:16 | es-index-cleaner-autoprov/1-install | starting test step 1-install logger.go:42: 07:41:16 | es-index-cleaner-autoprov/1-install | Jaeger:kuttl-test-classic-ghost/test-es-index-cleaner-with-prefix created logger.go:42: 07:41:51 | es-index-cleaner-autoprov/1-install | test step completed 1-install logger.go:42: 07:41:51 | es-index-cleaner-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:41:51 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE test-es-index-cleaner-with-prefix /dev/null] logger.go:42: 07:41:52 | es-index-cleaner-autoprov/2-report-spans | Warning: resource jaegers/test-es-index-cleaner-with-prefix is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:41:59 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c DAYS=5 ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:41:59 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:42:00 | es-index-cleaner-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:42:39 | es-index-cleaner-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 07:42:39 | es-index-cleaner-autoprov/3-install | starting test step 3-install logger.go:42: 07:42:39 | es-index-cleaner-autoprov/3-install | Jaeger:kuttl-test-classic-ghost/test-es-index-cleaner-with-prefix updated logger.go:42: 07:42:39 | es-index-cleaner-autoprov/3-install | test step completed 3-install logger.go:42: 07:42:39 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | starting test step 4-wait-es-index-cleaner logger.go:42: 07:42:39 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob test-es-index-cleaner-with-prefix-es-index-cleaner --namespace $NAMESPACE] logger.go:42: 07:42:40 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-12-25T07:42:40Z" level=debug msg="Checking if the test-es-index-cleaner-with-prefix-es-index-cleaner CronJob exists" logger.go:42: 07:42:40 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-12-25T07:42:40Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 07:42:40 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-12-25T07:42:40Z" level=warning msg="The BatchV1/Cronjob test-es-index-cleaner-with-prefix-es-index-cleaner was not found" logger.go:42: 07:42:40 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-12-25T07:42:40Z" level=debug msg="Found BatchV/Cronjobs:" logger.go:42: 07:42:50 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-12-25T07:42:50Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 07:42:50 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-12-25T07:42:50Z" level=info msg="Cronjob test-es-index-cleaner-with-prefix-es-index-cleaner found successfully" logger.go:42: 07:42:50 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-12-25T07:42:50Z" level=debug msg="Waiting for the next scheduled job from test-es-index-cleaner-with-prefix-es-index-cleaner cronjob" logger.go:42: 07:42:50 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-12-25T07:42:50Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 07:43:00 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-12-25T07:43:00Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 07:43:10 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-12-25T07:43:10Z" level=info msg="Job of owner test-es-index-cleaner-with-prefix-es-index-cleaner succeeded after test-es-index-cleaner-with-prefix-es-index-cleaner 20.029606586s" logger.go:42: 07:43:10 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | test step completed 4-wait-es-index-cleaner logger.go:42: 07:43:10 | es-index-cleaner-autoprov/5-install | starting test step 5-install logger.go:42: 07:43:10 | es-index-cleaner-autoprov/5-install | Jaeger:kuttl-test-classic-ghost/test-es-index-cleaner-with-prefix updated logger.go:42: 07:43:10 | es-index-cleaner-autoprov/5-install | test step completed 5-install logger.go:42: 07:43:10 | es-index-cleaner-autoprov/6-check-indices | starting test step 6-check-indices logger.go:42: 07:43:10 | es-index-cleaner-autoprov/6-check-indices | Job:kuttl-test-classic-ghost/00-check-indices created logger.go:42: 07:43:15 | es-index-cleaner-autoprov/6-check-indices | test step completed 6-check-indices logger.go:42: 07:43:15 | es-index-cleaner-autoprov | es-index-cleaner-autoprov events from ns kuttl-test-classic-ghost: logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclassicghosttestesindexclea-1-52b98b Binding Scheduled Successfully assigned kuttl-test-classic-ghost/elasticsearch-cdm-kuttltestclassicghosttestesindexclea-1-52b98b to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclassicghosttestesindexclea-1-52b98b AddedInterface Add eth0 [10.129.2.27/23] from ovn-kubernetes logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclassicghosttestesindexclea-1-52b98b.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclassicghosttestesindexclea-1-52b98b.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclassicghosttestesindexclea-1-52b98b.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclassicghosttestesindexclea-1-52b98b.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclassicghosttestesindexclea-1-52b98b.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclassicghosttestesindexclea-1-52b98b.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:22 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestclassicghosttestesindexclea-1-5f7f6f764c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestclassicghosttestesindexclea-1-52b98b replicaset-controller logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:22 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestclassicghosttestesindexclea-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestclassicghosttestesindexclea-1-5f7f6f764c to 1 deployment-controller logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:32 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestclassicghosttestesindexclea-1-52b98b.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:37 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestclassicghosttestesindexclea-1-52b98b.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-fdbdfd5f5-vwstw Binding Scheduled Successfully assigned kuttl-test-classic-ghost/test-es-index-cleaner-with-prefix-collector-fdbdfd5f5-vwstw to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-fdbdfd5f5-vwstw AddedInterface Add eth0 [10.128.2.42/23] from ovn-kubernetes logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-fdbdfd5f5-vwstw.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-fdbdfd5f5-vwstw.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-fdbdfd5f5-vwstw.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-collector-fdbdfd5f5 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-collector-fdbdfd5f5-vwstw replicaset-controller logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-collector ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-collector-fdbdfd5f5 to 1 deployment-controller logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b6b9869b7-zxgjb Binding Scheduled Successfully assigned kuttl-test-classic-ghost/test-es-index-cleaner-with-prefix-query-b6b9869b7-zxgjb to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b6b9869b7-zxgjb AddedInterface Add eth0 [10.131.0.29/23] from ovn-kubernetes logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b6b9869b7-zxgjb.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b6b9869b7-zxgjb.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b6b9869b7-zxgjb.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b6b9869b7-zxgjb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b6b9869b7-zxgjb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b6b9869b7-zxgjb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b6b9869b7-zxgjb.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b6b9869b7-zxgjb.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b6b9869b7-zxgjb.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-b6b9869b7 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-b6b9869b7-zxgjb replicaset-controller logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:49 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-b6b9869b7 to 1 deployment-controller logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:55 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b6b9869b7-zxgjb.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:55 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b6b9869b7-zxgjb.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:55 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b6b9869b7-zxgjb.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:55 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-b6b9869b7 SuccessfulDelete Deleted pod: test-es-index-cleaner-with-prefix-query-b6b9869b7-zxgjb replicaset-controller logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:55 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled down replica set test-es-index-cleaner-with-prefix-query-b6b9869b7 to 0 from 1 deployment-controller logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:56 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7c5744cd9c-jtgll Binding Scheduled Successfully assigned kuttl-test-classic-ghost/test-es-index-cleaner-with-prefix-query-7c5744cd9c-jtgll to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:56 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7c5744cd9c-jtgll AddedInterface Add eth0 [10.131.0.30/23] from ovn-kubernetes logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:56 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7c5744cd9c-jtgll.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:56 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7c5744cd9c-jtgll.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:56 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7c5744cd9c-jtgll.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:56 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7c5744cd9c-jtgll.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:56 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-7c5744cd9c SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-7c5744cd9c-jtgll replicaset-controller logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:56 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-7c5744cd9c to 1 deployment-controller logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7c5744cd9c-jtgll.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7c5744cd9c-jtgll.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7c5744cd9c-jtgll.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7c5744cd9c-jtgll.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:41:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7c5744cd9c-jtgll.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:42:00 +0000 UTC Normal Pod 00-report-span-scgm6 Binding Scheduled Successfully assigned kuttl-test-classic-ghost/00-report-span-scgm6 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:42:00 +0000 UTC Normal Pod 00-report-span-scgm6 AddedInterface Add eth0 [10.128.2.43/23] from ovn-kubernetes logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:42:00 +0000 UTC Normal Pod 00-report-span-scgm6.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:42:00 +0000 UTC Normal Pod 00-report-span-scgm6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:42:00 +0000 UTC Normal Pod 00-report-span-scgm6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:42:00 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-scgm6 job-controller logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:42:04 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:42:04 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:42:04 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:42:39 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:42:49 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:42:49 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-fdbdfd5f5-vwstw horizontal-pod-autoscaler logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:42:49 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:43:00 +0000 UTC Normal Job.batch test-es-index-cleaner-with-prefix-es-index-cleaner-28391503 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-es-index-cleaner-2839150zhl9b job-controller logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:43:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2839150zhl9b Binding Scheduled Successfully assigned kuttl-test-classic-ghost/test-es-index-cleaner-with-prefix-es-index-cleaner-2839150zhl9b to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:43:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2839150zhl9b AddedInterface Add eth0 [10.128.2.44/23] from ovn-kubernetes logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:43:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2839150zhl9b.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:e5bea9a64ae418869cfb556d70e0a586c21589aed8606b4ff5850780ff5bbbd6" kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:43:00 +0000 UTC Normal CronJob.batch test-es-index-cleaner-with-prefix-es-index-cleaner SuccessfulCreate Created job test-es-index-cleaner-with-prefix-es-index-cleaner-28391503 cronjob-controller logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:43:02 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2839150zhl9b.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:e5bea9a64ae418869cfb556d70e0a586c21589aed8606b4ff5850780ff5bbbd6" in 1.552s (1.552s including waiting) kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:43:02 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2839150zhl9b.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Created Created container test-es-index-cleaner-with-prefix-es-index-cleaner kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:43:02 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2839150zhl9b.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Started Started container test-es-index-cleaner-with-prefix-es-index-cleaner kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:43:05 +0000 UTC Normal Job.batch test-es-index-cleaner-with-prefix-es-index-cleaner-28391503 Completed Job completed job-controller logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:43:05 +0000 UTC Normal CronJob.batch test-es-index-cleaner-with-prefix-es-index-cleaner SawCompletedJob Saw completed job: test-es-index-cleaner-with-prefix-es-index-cleaner-28391503, status: Complete cronjob-controller logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:43:10 +0000 UTC Normal Pod 00-check-indices-4x2z5 Binding Scheduled Successfully assigned kuttl-test-classic-ghost/00-check-indices-4x2z5 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:43:10 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-4x2z5 job-controller logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:43:11 +0000 UTC Normal Pod 00-check-indices-4x2z5 AddedInterface Add eth0 [10.128.2.45/23] from ovn-kubernetes logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:43:11 +0000 UTC Normal Pod 00-check-indices-4x2z5.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:43:11 +0000 UTC Normal Pod 00-check-indices-4x2z5.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:43:11 +0000 UTC Normal Pod 00-check-indices-4x2z5.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:43:15 | es-index-cleaner-autoprov | 2023-12-25 07:43:14 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 07:43:15 | es-index-cleaner-autoprov | Deleting namespace: kuttl-test-classic-ghost === CONT kuttl/harness/es-from-aio-to-production logger.go:42: 07:43:22 | es-from-aio-to-production | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:43:22 | es-from-aio-to-production | Creating namespace: kuttl-test-busy-goldfish logger.go:42: 07:43:22 | es-from-aio-to-production/0-install | starting test step 0-install logger.go:42: 07:43:22 | es-from-aio-to-production/0-install | Jaeger:kuttl-test-busy-goldfish/my-jaeger created logger.go:42: 07:43:30 | es-from-aio-to-production/0-install | test step completed 0-install logger.go:42: 07:43:30 | es-from-aio-to-production/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:43:30 | es-from-aio-to-production/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:43:32 | es-from-aio-to-production/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:43:38 | es-from-aio-to-production/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:43:38 | es-from-aio-to-production/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:43:39 | es-from-aio-to-production/1-smoke-test | job.batch/report-span created logger.go:42: 07:43:39 | es-from-aio-to-production/1-smoke-test | job.batch/check-span created logger.go:42: 07:43:45 | es-from-aio-to-production/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:43:45 | es-from-aio-to-production/3-install | starting test step 3-install logger.go:42: 07:43:45 | es-from-aio-to-production/3-install | Jaeger:kuttl-test-busy-goldfish/my-jaeger updated logger.go:42: 07:44:19 | es-from-aio-to-production/3-install | test step completed 3-install logger.go:42: 07:44:19 | es-from-aio-to-production/4-smoke-test | starting test step 4-smoke-test logger.go:42: 07:44:19 | es-from-aio-to-production/4-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:44:27 | es-from-aio-to-production/4-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:44:27 | es-from-aio-to-production/4-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:44:28 | es-from-aio-to-production/4-smoke-test | job.batch/report-span unchanged logger.go:42: 07:44:28 | es-from-aio-to-production/4-smoke-test | job.batch/check-span unchanged logger.go:42: 07:44:28 | es-from-aio-to-production/4-smoke-test | test step completed 4-smoke-test logger.go:42: 07:44:28 | es-from-aio-to-production | es-from-aio-to-production events from ns kuttl-test-busy-goldfish: logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:25 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-f7cff5b7b to 1 deployment-controller logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:26 +0000 UTC Normal Pod my-jaeger-f7cff5b7b-gxjd7 Binding Scheduled Successfully assigned kuttl-test-busy-goldfish/my-jaeger-f7cff5b7b-gxjd7 to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:26 +0000 UTC Normal Pod my-jaeger-f7cff5b7b-gxjd7 AddedInterface Add eth0 [10.129.2.28/23] from ovn-kubernetes logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:26 +0000 UTC Normal Pod my-jaeger-f7cff5b7b-gxjd7.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:26 +0000 UTC Normal ReplicaSet.apps my-jaeger-f7cff5b7b SuccessfulCreate Created pod: my-jaeger-f7cff5b7b-gxjd7 replicaset-controller logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:28 +0000 UTC Normal Pod my-jaeger-f7cff5b7b-gxjd7.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" in 2.015s (2.015s including waiting) kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:28 +0000 UTC Normal Pod my-jaeger-f7cff5b7b-gxjd7.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:28 +0000 UTC Normal Pod my-jaeger-f7cff5b7b-gxjd7.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:28 +0000 UTC Normal Pod my-jaeger-f7cff5b7b-gxjd7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:28 +0000 UTC Normal Pod my-jaeger-f7cff5b7b-gxjd7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:28 +0000 UTC Normal Pod my-jaeger-f7cff5b7b-gxjd7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:34 +0000 UTC Normal Pod my-jaeger-f7cff5b7b-gxjd7.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:34 +0000 UTC Normal Pod my-jaeger-f7cff5b7b-gxjd7.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:34 +0000 UTC Normal ReplicaSet.apps my-jaeger-f7cff5b7b SuccessfulDelete Deleted pod: my-jaeger-f7cff5b7b-gxjd7 replicaset-controller logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:34 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-f7cff5b7b to 0 from 1 deployment-controller logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:35 +0000 UTC Normal Pod my-jaeger-6495db6d75-hpmc9 Binding Scheduled Successfully assigned kuttl-test-busy-goldfish/my-jaeger-6495db6d75-hpmc9 to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:35 +0000 UTC Normal Pod my-jaeger-6495db6d75-hpmc9 AddedInterface Add eth0 [10.129.2.29/23] from ovn-kubernetes logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:35 +0000 UTC Normal Pod my-jaeger-6495db6d75-hpmc9.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:35 +0000 UTC Normal Pod my-jaeger-6495db6d75-hpmc9.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:35 +0000 UTC Normal Pod my-jaeger-6495db6d75-hpmc9.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:35 +0000 UTC Normal Pod my-jaeger-6495db6d75-hpmc9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:35 +0000 UTC Normal Pod my-jaeger-6495db6d75-hpmc9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:35 +0000 UTC Normal Pod my-jaeger-6495db6d75-hpmc9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:35 +0000 UTC Normal ReplicaSet.apps my-jaeger-6495db6d75 SuccessfulCreate Created pod: my-jaeger-6495db6d75-hpmc9 replicaset-controller logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:35 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6495db6d75 to 1 deployment-controller logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:39 +0000 UTC Normal Pod check-span-98x79 Binding Scheduled Successfully assigned kuttl-test-busy-goldfish/check-span-98x79 to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:39 +0000 UTC Normal Pod check-span-98x79 AddedInterface Add eth0 [10.131.0.31/23] from ovn-kubernetes logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:39 +0000 UTC Normal Pod check-span-98x79.spec.containers{asserts-container} Pulling Pulling image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:39 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-98x79 job-controller logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:39 +0000 UTC Normal Pod report-span-mmrt2 Binding Scheduled Successfully assigned kuttl-test-busy-goldfish/report-span-mmrt2 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:39 +0000 UTC Normal Pod report-span-mmrt2 AddedInterface Add eth0 [10.128.2.46/23] from ovn-kubernetes logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:39 +0000 UTC Normal Pod report-span-mmrt2.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:39 +0000 UTC Normal Pod report-span-mmrt2.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:39 +0000 UTC Normal Pod report-span-mmrt2.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:39 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-mmrt2 job-controller logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:42 +0000 UTC Normal Pod check-span-98x79.spec.containers{asserts-container} Pulled Successfully pulled image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" in 2.321s (2.321s including waiting) kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:42 +0000 UTC Normal Pod check-span-98x79.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:42 +0000 UTC Normal Pod check-span-98x79.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:44 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:50 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestbusygoldfishmyjaeger-1-64779db9f5 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestbusygoldfishmyjaeger-1-64779db9kktjh replicaset-controller logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusygoldfishmyjaeger-1-64779db9kktjh Binding Scheduled Successfully assigned kuttl-test-busy-goldfish/elasticsearch-cdm-kuttltestbusygoldfishmyjaeger-1-64779db9kktjh to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusygoldfishmyjaeger-1-64779db9kktjh AddedInterface Add eth0 [10.129.2.30/23] from ovn-kubernetes logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusygoldfishmyjaeger-1-64779db9kktjh.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusygoldfishmyjaeger-1-64779db9kktjh.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusygoldfishmyjaeger-1-64779db9kktjh.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusygoldfishmyjaeger-1-64779db9kktjh.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusygoldfishmyjaeger-1-64779db9kktjh.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusygoldfishmyjaeger-1-64779db9kktjh.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:43:50 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestbusygoldfishmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestbusygoldfishmyjaeger-1-64779db9f5 to 1 deployment-controller logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:00 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestbusygoldfishmyjaeger-1-64779db9kktjh.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:05 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestbusygoldfishmyjaeger-1-64779db9kktjh.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:12 +0000 UTC Normal Job.batch report-span Completed Job completed job-controller logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:16 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-74b8877b7c to 1 deployment-controller logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:16 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-6979646c8f to 1 deployment-controller logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal Pod my-jaeger-6495db6d75-hpmc9.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal Pod my-jaeger-6495db6d75-hpmc9.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal Pod my-jaeger-collector-74b8877b7c-8ktgf Binding Scheduled Successfully assigned kuttl-test-busy-goldfish/my-jaeger-collector-74b8877b7c-8ktgf to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal Pod my-jaeger-collector-74b8877b7c-8ktgf AddedInterface Add eth0 [10.128.2.47/23] from ovn-kubernetes logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal Pod my-jaeger-collector-74b8877b7c-8ktgf.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal Pod my-jaeger-collector-74b8877b7c-8ktgf.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal Pod my-jaeger-collector-74b8877b7c-8ktgf.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-74b8877b7c SuccessfulCreate Created pod: my-jaeger-collector-74b8877b7c-8ktgf replicaset-controller logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal Pod my-jaeger-query-6979646c8f-7g74t Binding Scheduled Successfully assigned kuttl-test-busy-goldfish/my-jaeger-query-6979646c8f-7g74t to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal Pod my-jaeger-query-6979646c8f-7g74t AddedInterface Add eth0 [10.131.0.32/23] from ovn-kubernetes logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal Pod my-jaeger-query-6979646c8f-7g74t.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal Pod my-jaeger-query-6979646c8f-7g74t.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal Pod my-jaeger-query-6979646c8f-7g74t.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal Pod my-jaeger-query-6979646c8f-7g74t.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal Pod my-jaeger-query-6979646c8f-7g74t.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal Pod my-jaeger-query-6979646c8f-7g74t.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal Pod my-jaeger-query-6979646c8f-7g74t.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal Pod my-jaeger-query-6979646c8f-7g74t.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal Pod my-jaeger-query-6979646c8f-7g74t.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:44:28 | es-from-aio-to-production | 2023-12-25 07:44:17 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-6979646c8f SuccessfulCreate Created pod: my-jaeger-query-6979646c8f-7g74t replicaset-controller logger.go:42: 07:44:28 | es-from-aio-to-production | Deleting namespace: kuttl-test-busy-goldfish === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (669.42s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.32s) --- PASS: kuttl/harness/es-multiinstance (111.05s) --- PASS: kuttl/harness/es-simple-prod (6.31s) --- PASS: kuttl/harness/es-rollover-autoprov (238.42s) --- PASS: kuttl/harness/es-increasing-replicas (108.21s) --- PASS: kuttl/harness/es-index-cleaner-autoprov (126.07s) --- PASS: kuttl/harness/es-from-aio-to-production (72.89s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name elasticsearch --report --output /logs/artifacts/elasticsearch.xml ./artifacts/kuttl-report.xml time="2023-12-25T07:44:36Z" level=debug msg="Setting a new name for the test suites" time="2023-12-25T07:44:36Z" level=debug msg="Removing 'artifacts' TestCase" time="2023-12-25T07:44:36Z" level=debug msg="normalizing test case names" time="2023-12-25T07:44:36Z" level=debug msg="elasticsearch/artifacts -> elasticsearch_artifacts" time="2023-12-25T07:44:36Z" level=debug msg="elasticsearch/es-multiinstance -> elasticsearch_es_multiinstance" time="2023-12-25T07:44:36Z" level=debug msg="elasticsearch/es-simple-prod -> elasticsearch_es_simple_prod" time="2023-12-25T07:44:36Z" level=debug msg="elasticsearch/es-rollover-autoprov -> elasticsearch_es_rollover_autoprov" time="2023-12-25T07:44:36Z" level=debug msg="elasticsearch/es-increasing-replicas -> elasticsearch_es_increasing_replicas" time="2023-12-25T07:44:36Z" level=debug msg="elasticsearch/es-index-cleaner-autoprov -> elasticsearch_es_index_cleaner_autoprov" time="2023-12-25T07:44:36Z" level=debug msg="elasticsearch/es-from-aio-to-production -> elasticsearch_es_from_aio_to_production" +-----------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------+--------+ | elasticsearch_artifacts | passed | | elasticsearch_es_multiinstance | passed | | elasticsearch_es_simple_prod | passed | | elasticsearch_es_rollover_autoprov | passed | | elasticsearch_es_increasing_replicas | passed | | elasticsearch_es_index_cleaner_autoprov | passed | | elasticsearch_es_from_aio_to_production | passed | +-----------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh examples false true + '[' 3 -ne 3 ']' + test_suite_name=examples + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/examples.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-examples make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ VERTX_IMG=jaegertracing/vertx-create-span:operator-e2e-tests \ ./tests/e2e/examples/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-12-22-165425 True False 18m Cluster version is 4.15.0-0.nightly-2023-12-22-165425' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-12-22-165425 True False 18m Cluster version is 4.15.0-0.nightly-2023-12-22-165425' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/examples/render.sh ++ export SUITE_DIR=./tests/e2e/examples ++ SUITE_DIR=./tests/e2e/examples ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/examples ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test examples-agent-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-agent-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-with-priority-class\e[0m' Rendering files for test examples-agent-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + mkdir -p examples-agent-with-priority-class + cd examples-agent-with-priority-class + example_name=agent-with-priority-class + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + '[' true '!=' true ']' + render_install_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + install_file=./02-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/agent-with-priority-class.yaml -o ./02-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./02-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./02-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./02-install.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + JAEGER_NAME=agent-as-daemonset + local jaeger_strategy ++ get_jaeger_strategy ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./02-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./02-install.yaml ++ strategy=DaemonSet ++ '[' DaemonSet = null ']' ++ echo DaemonSet ++ return 0 + jaeger_strategy=DaemonSet + '[' DaemonSet = DaemonSet ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./02-assert.yaml + render_smoke_test_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + jaeger_name=agent-as-daemonset + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test agent-as-daemonset true 02 + '[' 3 -ne 3 ']' + jaeger=agent-as-daemonset + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-all-in-one-with-options + '[' 1 -ne 1 ']' + test_name=examples-all-in-one-with-options + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-all-in-one-with-options' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-all-in-one-with-options\e[0m' Rendering files for test examples-all-in-one-with-options + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-with-priority-class + '[' examples-agent-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-all-in-one-with-options + cd examples-all-in-one-with-options + example_name=all-in-one-with-options + render_install_example all-in-one-with-options 00 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/all-in-one-with-options.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + JAEGER_NAME=my-jaeger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.metadata.name="my-jaeger"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i 'del(.spec.allInOne.image)' ./00-install.yaml + render_smoke_test_example all-in-one-with-options 01 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + jaeger_name=my-jaeger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + sed -i s~my-jaeger-query:443~my-jaeger-query:443/jaeger~gi ./01-smoke-test.yaml + start_test examples-business-application-injected-sidecar + '[' 1 -ne 1 ']' + test_name=examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-business-application-injected-sidecar' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-business-application-injected-sidecar\e[0m' Rendering files for test examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-all-in-one-with-options + '[' examples-all-in-one-with-options '!=' _build ']' + cd .. + mkdir -p examples-business-application-injected-sidecar + cd examples-business-application-injected-sidecar + example_name=simplest + cp /tmp/jaeger-tests/examples/business-application-injected-sidecar.yaml ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].image=strenv(VERTX_IMG)' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.path="/"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.port=8080' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.initialDelaySeconds=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.failureThreshold=3' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.periodSeconds=10' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.successThreshold=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.timeoutSeconds=1' ./00-install.yaml + render_install_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example simplest 02 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 02 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-collector-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-collector-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-collector-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-collector-with-priority-class\e[0m' Rendering files for test examples-collector-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-business-application-injected-sidecar + '[' examples-business-application-injected-sidecar '!=' _build ']' + cd .. + mkdir -p examples-collector-with-priority-class + cd examples-collector-with-priority-class + example_name=collector-with-priority-class + render_install_example collector-with-priority-class 00 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/collector-with-priority-class.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + JAEGER_NAME=collector-with-high-priority + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example collector-with-priority-class 01 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + jaeger_name=collector-with-high-priority + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test collector-with-high-priority true 01 + '[' 3 -ne 3 ']' + jaeger=collector-with-high-priority + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + export JAEGER_NAME=collector-with-high-priority + JAEGER_NAME=collector-with-high-priority + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-service-types + '[' 1 -ne 1 ']' + test_name=examples-service-types + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-service-types' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-service-types\e[0m' Rendering files for test examples-service-types + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-collector-with-priority-class + '[' examples-collector-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-service-types + cd examples-service-types + example_name=service-types + render_install_example service-types 00 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/service-types.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + JAEGER_NAME=service-types + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example service-types 01 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/service-types.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/service-types.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/service-types.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/service-types.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + jaeger_name=service-types + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test service-types true 01 + '[' 3 -ne 3 ']' + jaeger=service-types + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + export JAEGER_NAME=service-types + JAEGER_NAME=service-types + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod + '[' 1 -ne 1 ']' + test_name=examples-simple-prod + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod\e[0m' Rendering files for test examples-simple-prod + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-service-types + '[' examples-service-types '!=' _build ']' + cd .. + mkdir -p examples-simple-prod + cd examples-simple-prod + example_name=simple-prod + render_install_example simple-prod 01 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod 02 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod-with-volumes + '[' 1 -ne 1 ']' + test_name=examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod-with-volumes' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod-with-volumes\e[0m' Rendering files for test examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod + '[' examples-simple-prod '!=' _build ']' + cd .. + mkdir -p examples-simple-prod-with-volumes + cd examples-simple-prod-with-volumes + example_name=simple-prod-with-volumes + render_install_example simple-prod-with-volumes 01 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod-with-volumes 02 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + /tmp/jaeger-tests/bin/gomplate -f ./03-check-volume.yaml.template -o 03-check-volume.yaml + start_test examples-simplest + '[' 1 -ne 1 ']' + test_name=examples-simplest + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simplest' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simplest\e[0m' Rendering files for test examples-simplest + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod-with-volumes + '[' examples-simple-prod-with-volumes '!=' _build ']' + cd .. + mkdir -p examples-simplest + cd examples-simplest + example_name=simplest + render_install_example simplest 00 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 01 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger + '[' 1 -ne 1 ']' + test_name=examples-with-badger + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger\e[0m' Rendering files for test examples-with-badger + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simplest + '[' examples-simplest '!=' _build ']' + cd .. + mkdir -p examples-with-badger + cd examples-with-badger + example_name=with-badger + render_install_example with-badger 00 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + JAEGER_NAME=with-badger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger 01 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + jaeger_name=with-badger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + export JAEGER_NAME=with-badger + JAEGER_NAME=with-badger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger-and-volume + '[' 1 -ne 1 ']' + test_name=examples-with-badger-and-volume + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger-and-volume' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger-and-volume\e[0m' Rendering files for test examples-with-badger-and-volume + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger + '[' examples-with-badger '!=' _build ']' + cd .. + mkdir -p examples-with-badger-and-volume + cd examples-with-badger-and-volume + example_name=with-badger-and-volume + render_install_example with-badger-and-volume 00 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger-and-volume.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + JAEGER_NAME=with-badger-and-volume + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger-and-volume 01 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + jaeger_name=with-badger-and-volume + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger-and-volume true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger-and-volume + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + export JAEGER_NAME=with-badger-and-volume + JAEGER_NAME=with-badger-and-volume + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-cassandra + '[' 1 -ne 1 ']' + test_name=examples-with-cassandra + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-cassandra' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-cassandra\e[0m' Rendering files for test examples-with-cassandra + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger-and-volume + '[' examples-with-badger-and-volume '!=' _build ']' + cd .. + mkdir -p examples-with-cassandra + cd examples-with-cassandra + example_name=with-cassandra + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-cassandra 01 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-cassandra.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + JAEGER_NAME=with-cassandra + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-cassandra 02 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-cassandra.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-cassandra.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + jaeger_name=with-cassandra + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-cassandra true 02 + '[' 3 -ne 3 ']' + jaeger=with-cassandra + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + export JAEGER_NAME=with-cassandra + JAEGER_NAME=with-cassandra + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-sampling + '[' 1 -ne 1 ']' + test_name=examples-with-sampling + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-sampling' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-sampling\e[0m' Rendering files for test examples-with-sampling + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-cassandra + '[' examples-with-cassandra '!=' _build ']' + cd .. + mkdir -p examples-with-sampling + cd examples-with-sampling + export example_name=with-sampling + example_name=with-sampling + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-sampling 01 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-sampling.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + JAEGER_NAME=with-sampling + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-sampling 02 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-sampling.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-sampling.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + jaeger_name=with-sampling + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-sampling true 02 + '[' 3 -ne 3 ']' + jaeger=with-sampling + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + export JAEGER_NAME=with-sampling + JAEGER_NAME=with-sampling + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-agent-as-daemonset + '[' 1 -ne 1 ']' + test_name=examples-agent-as-daemonset + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-as-daemonset' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-as-daemonset\e[0m' Rendering files for test examples-agent-as-daemonset + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-sampling + '[' examples-with-sampling '!=' _build ']' + cd .. + mkdir -p examples-agent-as-daemonset + cd examples-agent-as-daemonset + '[' true = true ']' + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/agent-as-daemonset.yaml -o 02-install.yaml + '[' true = true ']' + start_test examples-openshift-with-htpasswd + '[' 1 -ne 1 ']' + test_name=examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-openshift-with-htpasswd' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-openshift-with-htpasswd\e[0m' Rendering files for test examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-as-daemonset + '[' examples-agent-as-daemonset '!=' _build ']' + cd .. + mkdir -p examples-openshift-with-htpasswd + cd examples-openshift-with-htpasswd + export JAEGER_NAME=with-htpasswd + JAEGER_NAME=with-htpasswd + export JAEGER_USERNAME=awesomeuser + JAEGER_USERNAME=awesomeuser + export JAEGER_PASSWORD=awesomepassword + JAEGER_PASSWORD=awesomepassword + export 'JAEGER_USER_PASSWORD_HASH=awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' + JAEGER_USER_PASSWORD_HASH='awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ echo 'awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ base64 + SECRET=YXdlc29tZXVzZXI6e1NIQX11VWRxUFZVeXFOQm1FUlUwUXhqM0tGYVpuanc9Cg== + /tmp/jaeger-tests/bin/gomplate -f ./00-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/with-htpasswd.yaml -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + export 'GET_URL_COMMAND=kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + export 'URL=https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + INSECURE=true + JAEGER_USERNAME= + JAEGER_PASSWORD= + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-unsecured.yaml + JAEGER_USERNAME=wronguser + JAEGER_PASSWORD=wrongpassword + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-check-unauthorized.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./04-check-authorized.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running examples E2E tests' Running examples E2E tests + cd tests/e2e/examples/_build + set +e + KUBECONFIG=/tmp/kubeconfig-1441553849 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 15 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/examples-agent-as-daemonset === PAUSE kuttl/harness/examples-agent-as-daemonset === RUN kuttl/harness/examples-agent-with-priority-class === PAUSE kuttl/harness/examples-agent-with-priority-class === RUN kuttl/harness/examples-all-in-one-with-options === PAUSE kuttl/harness/examples-all-in-one-with-options === RUN kuttl/harness/examples-business-application-injected-sidecar === PAUSE kuttl/harness/examples-business-application-injected-sidecar === RUN kuttl/harness/examples-collector-with-priority-class === PAUSE kuttl/harness/examples-collector-with-priority-class === RUN kuttl/harness/examples-openshift-with-htpasswd === PAUSE kuttl/harness/examples-openshift-with-htpasswd === RUN kuttl/harness/examples-service-types === PAUSE kuttl/harness/examples-service-types === RUN kuttl/harness/examples-simple-prod === PAUSE kuttl/harness/examples-simple-prod === RUN kuttl/harness/examples-simple-prod-with-volumes === PAUSE kuttl/harness/examples-simple-prod-with-volumes === RUN kuttl/harness/examples-simplest === PAUSE kuttl/harness/examples-simplest === RUN kuttl/harness/examples-with-badger === PAUSE kuttl/harness/examples-with-badger === RUN kuttl/harness/examples-with-badger-and-volume === PAUSE kuttl/harness/examples-with-badger-and-volume === RUN kuttl/harness/examples-with-cassandra === PAUSE kuttl/harness/examples-with-cassandra === RUN kuttl/harness/examples-with-sampling === PAUSE kuttl/harness/examples-with-sampling === CONT kuttl/harness/artifacts logger.go:42: 07:45:09 | artifacts | Creating namespace: kuttl-test-precise-buck logger.go:42: 07:45:09 | artifacts | artifacts events from ns kuttl-test-precise-buck: logger.go:42: 07:45:09 | artifacts | Deleting namespace: kuttl-test-precise-buck === CONT kuttl/harness/examples-simple-prod logger.go:42: 07:45:15 | examples-simple-prod | Creating namespace: kuttl-test-adapted-albacore logger.go:42: 07:45:15 | examples-simple-prod/1-install | starting test step 1-install logger.go:42: 07:45:15 | examples-simple-prod/1-install | Jaeger:kuttl-test-adapted-albacore/simple-prod created logger.go:42: 07:45:51 | examples-simple-prod/1-install | test step completed 1-install logger.go:42: 07:45:51 | examples-simple-prod/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:45:51 | examples-simple-prod/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:45:53 | examples-simple-prod/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:45:59 | examples-simple-prod/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:46:00 | examples-simple-prod/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:46:00 | examples-simple-prod/2-smoke-test | job.batch/report-span created logger.go:42: 07:46:00 | examples-simple-prod/2-smoke-test | job.batch/check-span created logger.go:42: 07:46:11 | examples-simple-prod/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:46:11 | examples-simple-prod | examples-simple-prod events from ns kuttl-test-adapted-albacore: logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestadaptedalbacoresimpleprod-1-65d7rgxd Binding Scheduled Successfully assigned kuttl-test-adapted-albacore/elasticsearch-cdm-kuttltestadaptedalbacoresimpleprod-1-65d7rgxd to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:20 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestadaptedalbacoresimpleprod-1-65db87f45c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestadaptedalbacoresimpleprod-1-65d7rgxd replicaset-controller logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:20 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestadaptedalbacoresimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestadaptedalbacoresimpleprod-1-65db87f45c to 1 deployment-controller logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestadaptedalbacoresimpleprod-1-65d7rgxd AddedInterface Add eth0 [10.129.2.32/23] from ovn-kubernetes logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestadaptedalbacoresimpleprod-1-65d7rgxd.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestadaptedalbacoresimpleprod-1-65d7rgxd.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestadaptedalbacoresimpleprod-1-65d7rgxd.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestadaptedalbacoresimpleprod-1-65d7rgxd.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestadaptedalbacoresimpleprod-1-65d7rgxd.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestadaptedalbacoresimpleprod-1-65d7rgxd.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:31 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestadaptedalbacoresimpleprod-1-65d7rgxd.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:36 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestadaptedalbacoresimpleprod-1-65d7rgxd.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:47 +0000 UTC Normal Pod simple-prod-collector-58c68449c8-w8lmk Binding Scheduled Successfully assigned kuttl-test-adapted-albacore/simple-prod-collector-58c68449c8-w8lmk to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:47 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-58c68449c8 SuccessfulCreate Created pod: simple-prod-collector-58c68449c8-w8lmk replicaset-controller logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:47 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-58c68449c8 to 1 deployment-controller logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:47 +0000 UTC Normal Pod simple-prod-query-57c7b64bc5-x26qr Binding Scheduled Successfully assigned kuttl-test-adapted-albacore/simple-prod-query-57c7b64bc5-x26qr to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:47 +0000 UTC Normal ReplicaSet.apps simple-prod-query-57c7b64bc5 SuccessfulCreate Created pod: simple-prod-query-57c7b64bc5-x26qr replicaset-controller logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:47 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-57c7b64bc5 to 1 deployment-controller logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:48 +0000 UTC Warning Pod simple-prod-query-57c7b64bc5-x26qr FailedMount MountVolume.SetUp failed for volume "simple-prod-ui-oauth-proxy-tls" : secret "simple-prod-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:48 +0000 UTC Normal Pod simple-prod-query-57c7b64bc5-x26qr AddedInterface Add eth0 [10.131.0.33/23] from ovn-kubernetes logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:48 +0000 UTC Normal Pod simple-prod-query-57c7b64bc5-x26qr.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:48 +0000 UTC Normal Pod simple-prod-query-57c7b64bc5-x26qr.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:48 +0000 UTC Normal Pod simple-prod-query-57c7b64bc5-x26qr.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:48 +0000 UTC Normal Pod simple-prod-query-57c7b64bc5-x26qr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:48 +0000 UTC Normal Pod simple-prod-query-57c7b64bc5-x26qr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:49 +0000 UTC Warning Pod simple-prod-collector-58c68449c8-w8lmk FailedMount MountVolume.SetUp failed for volume "certs" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:49 +0000 UTC Warning Pod simple-prod-collector-58c68449c8-w8lmk FailedMount MountVolume.SetUp failed for volume "simple-prod-sampling-configuration-volume" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:49 +0000 UTC Warning Pod simple-prod-collector-58c68449c8-w8lmk FailedMount MountVolume.SetUp failed for volume "simple-prod-collector-tls-config-volume" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:49 +0000 UTC Normal Pod simple-prod-collector-58c68449c8-w8lmk AddedInterface Add eth0 [10.128.2.48/23] from ovn-kubernetes logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:49 +0000 UTC Normal Pod simple-prod-collector-58c68449c8-w8lmk.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:49 +0000 UTC Normal Pod simple-prod-collector-58c68449c8-w8lmk.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:49 +0000 UTC Normal Pod simple-prod-collector-58c68449c8-w8lmk.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:49 +0000 UTC Normal Pod simple-prod-query-57c7b64bc5-x26qr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:49 +0000 UTC Normal Pod simple-prod-query-57c7b64bc5-x26qr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:49 +0000 UTC Normal Pod simple-prod-query-57c7b64bc5-x26qr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:49 +0000 UTC Normal Pod simple-prod-query-57c7b64bc5-x26qr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:55 +0000 UTC Normal Pod simple-prod-query-57c7b64bc5-x26qr.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:55 +0000 UTC Normal Pod simple-prod-query-57c7b64bc5-x26qr.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:55 +0000 UTC Normal Pod simple-prod-query-57c7b64bc5-x26qr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:55 +0000 UTC Normal ReplicaSet.apps simple-prod-query-57c7b64bc5 SuccessfulDelete Deleted pod: simple-prod-query-57c7b64bc5-x26qr replicaset-controller logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:55 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-57c7b64bc5 to 0 from 1 deployment-controller logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:55 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-77fdfb86f8 to 1 deployment-controller logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:56 +0000 UTC Normal Pod simple-prod-query-77fdfb86f8-mhfjf Binding Scheduled Successfully assigned kuttl-test-adapted-albacore/simple-prod-query-77fdfb86f8-mhfjf to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:56 +0000 UTC Normal Pod simple-prod-query-77fdfb86f8-mhfjf AddedInterface Add eth0 [10.131.0.34/23] from ovn-kubernetes logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:56 +0000 UTC Normal Pod simple-prod-query-77fdfb86f8-mhfjf.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:56 +0000 UTC Normal Pod simple-prod-query-77fdfb86f8-mhfjf.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:56 +0000 UTC Normal Pod simple-prod-query-77fdfb86f8-mhfjf.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:56 +0000 UTC Normal Pod simple-prod-query-77fdfb86f8-mhfjf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:56 +0000 UTC Normal Pod simple-prod-query-77fdfb86f8-mhfjf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:56 +0000 UTC Normal Pod simple-prod-query-77fdfb86f8-mhfjf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:56 +0000 UTC Normal Pod simple-prod-query-77fdfb86f8-mhfjf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:56 +0000 UTC Normal Pod simple-prod-query-77fdfb86f8-mhfjf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:56 +0000 UTC Normal Pod simple-prod-query-77fdfb86f8-mhfjf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:45:56 +0000 UTC Normal ReplicaSet.apps simple-prod-query-77fdfb86f8 SuccessfulCreate Created pod: simple-prod-query-77fdfb86f8-mhfjf replicaset-controller logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:46:00 +0000 UTC Normal Pod check-span-z9mk8 Binding Scheduled Successfully assigned kuttl-test-adapted-albacore/check-span-z9mk8 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:46:00 +0000 UTC Normal Pod check-span-z9mk8 AddedInterface Add eth0 [10.128.2.50/23] from ovn-kubernetes logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:46:00 +0000 UTC Normal Pod check-span-z9mk8.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:46:00 +0000 UTC Normal Pod check-span-z9mk8.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:46:00 +0000 UTC Normal Pod check-span-z9mk8.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:46:00 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-z9mk8 job-controller logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:46:00 +0000 UTC Normal Pod report-span-mdk8k Binding Scheduled Successfully assigned kuttl-test-adapted-albacore/report-span-mdk8k to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:46:00 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-mdk8k job-controller logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:46:01 +0000 UTC Normal Pod report-span-mdk8k AddedInterface Add eth0 [10.128.2.49/23] from ovn-kubernetes logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:46:01 +0000 UTC Normal Pod report-span-mdk8k.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:46:01 +0000 UTC Normal Pod report-span-mdk8k.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:46:01 +0000 UTC Normal Pod report-span-mdk8k.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:46:02 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:46:02 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:46:02 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:46:11 | examples-simple-prod | 2023-12-25 07:46:11 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:46:11 | examples-simple-prod | Deleting namespace: kuttl-test-adapted-albacore === CONT kuttl/harness/examples-with-sampling logger.go:42: 07:46:24 | examples-with-sampling | Creating namespace: kuttl-test-destined-lacewing logger.go:42: 07:46:24 | examples-with-sampling/0-install | starting test step 0-install logger.go:42: 07:46:24 | examples-with-sampling/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:46:24 | examples-with-sampling/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:46:24 | examples-with-sampling/0-install | >>>> Creating namespace kuttl-test-destined-lacewing logger.go:42: 07:46:24 | examples-with-sampling/0-install | kubectl create namespace kuttl-test-destined-lacewing 2>&1 | grep -v "already exists" || true logger.go:42: 07:46:24 | examples-with-sampling/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-destined-lacewing 2>&1 | grep -v "already exists" || true logger.go:42: 07:46:25 | examples-with-sampling/0-install | service/cassandra created logger.go:42: 07:46:25 | examples-with-sampling/0-install | statefulset.apps/cassandra created logger.go:42: 07:46:25 | examples-with-sampling/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:46:36 | examples-with-sampling/0-install | test step completed 0-install logger.go:42: 07:46:36 | examples-with-sampling/1-install | starting test step 1-install logger.go:42: 07:46:36 | examples-with-sampling/1-install | Jaeger:kuttl-test-destined-lacewing/with-sampling created logger.go:42: 07:46:42 | examples-with-sampling/1-install | test step completed 1-install logger.go:42: 07:46:42 | examples-with-sampling/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:46:42 | examples-with-sampling/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-sampling /dev/null] logger.go:42: 07:46:44 | examples-with-sampling/2-smoke-test | Warning: resource jaegers/with-sampling is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:46:50 | examples-with-sampling/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:46:51 | examples-with-sampling/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:46:51 | examples-with-sampling/2-smoke-test | job.batch/report-span created logger.go:42: 07:46:51 | examples-with-sampling/2-smoke-test | job.batch/check-span created logger.go:42: 07:47:02 | examples-with-sampling/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:47:02 | examples-with-sampling/3- | starting test step 3- logger.go:42: 07:47:02 | examples-with-sampling/3- | test step completed 3- logger.go:42: 07:47:02 | examples-with-sampling | examples-with-sampling events from ns kuttl-test-destined-lacewing: logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:25 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-destined-lacewing/cassandra-0 to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:25 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.129.2.33/23] from ovn-kubernetes logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:25 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:25 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:30 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 4.222s (4.222s including waiting) kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:30 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:30 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:30 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-destined-lacewing/cassandra-1 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:30 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:31 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.128.2.51/23] from ovn-kubernetes logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:31 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:35 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 4.067s (4.067s including waiting) kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:35 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:35 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:40 +0000 UTC Normal Pod with-sampling-58846657d7-rwjxr Binding Scheduled Successfully assigned kuttl-test-destined-lacewing/with-sampling-58846657d7-rwjxr to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:40 +0000 UTC Normal Pod with-sampling-58846657d7-rwjxr AddedInterface Add eth0 [10.129.2.34/23] from ovn-kubernetes logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:40 +0000 UTC Normal Pod with-sampling-58846657d7-rwjxr.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:40 +0000 UTC Normal Pod with-sampling-58846657d7-rwjxr.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:40 +0000 UTC Normal Pod with-sampling-58846657d7-rwjxr.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:40 +0000 UTC Normal Pod with-sampling-58846657d7-rwjxr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:40 +0000 UTC Normal Pod with-sampling-58846657d7-rwjxr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:40 +0000 UTC Normal Pod with-sampling-58846657d7-rwjxr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:40 +0000 UTC Normal ReplicaSet.apps with-sampling-58846657d7 SuccessfulCreate Created pod: with-sampling-58846657d7-rwjxr replicaset-controller logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:40 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-58846657d7 to 1 deployment-controller logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:46 +0000 UTC Normal Pod with-sampling-58846657d7-rwjxr.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:46 +0000 UTC Normal Pod with-sampling-58846657d7-rwjxr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:46 +0000 UTC Normal ReplicaSet.apps with-sampling-58846657d7 SuccessfulDelete Deleted pod: with-sampling-58846657d7-rwjxr replicaset-controller logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:46 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled down replica set with-sampling-58846657d7 to 0 from 1 deployment-controller logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:47 +0000 UTC Normal Pod with-sampling-5bb65f996f-qxxkk Binding Scheduled Successfully assigned kuttl-test-destined-lacewing/with-sampling-5bb65f996f-qxxkk to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:47 +0000 UTC Normal ReplicaSet.apps with-sampling-5bb65f996f SuccessfulCreate Created pod: with-sampling-5bb65f996f-qxxkk replicaset-controller logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:47 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-5bb65f996f to 1 deployment-controller logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:48 +0000 UTC Normal Pod with-sampling-5bb65f996f-qxxkk AddedInterface Add eth0 [10.129.2.35/23] from ovn-kubernetes logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:48 +0000 UTC Normal Pod with-sampling-5bb65f996f-qxxkk.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:48 +0000 UTC Normal Pod with-sampling-5bb65f996f-qxxkk.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:48 +0000 UTC Normal Pod with-sampling-5bb65f996f-qxxkk.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:48 +0000 UTC Normal Pod with-sampling-5bb65f996f-qxxkk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:48 +0000 UTC Normal Pod with-sampling-5bb65f996f-qxxkk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:48 +0000 UTC Normal Pod with-sampling-5bb65f996f-qxxkk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:51 +0000 UTC Normal Pod check-span-nfcrq Binding Scheduled Successfully assigned kuttl-test-destined-lacewing/check-span-nfcrq to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:51 +0000 UTC Normal Pod check-span-nfcrq AddedInterface Add eth0 [10.131.0.36/23] from ovn-kubernetes logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:51 +0000 UTC Normal Pod check-span-nfcrq.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:51 +0000 UTC Normal Pod check-span-nfcrq.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:51 +0000 UTC Normal Pod check-span-nfcrq.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:51 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-nfcrq job-controller logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:51 +0000 UTC Normal Pod report-span-k4pcd Binding Scheduled Successfully assigned kuttl-test-destined-lacewing/report-span-k4pcd to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:51 +0000 UTC Normal Pod report-span-k4pcd AddedInterface Add eth0 [10.131.0.35/23] from ovn-kubernetes logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:51 +0000 UTC Normal Pod report-span-k4pcd.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:51 +0000 UTC Normal Pod report-span-k4pcd.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:51 +0000 UTC Normal Pod report-span-k4pcd.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:46:51 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-k4pcd job-controller logger.go:42: 07:47:02 | examples-with-sampling | 2023-12-25 07:47:02 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:47:02 | examples-with-sampling | Deleting namespace: kuttl-test-destined-lacewing === CONT kuttl/harness/examples-with-cassandra logger.go:42: 07:47:21 | examples-with-cassandra | Creating namespace: kuttl-test-moved-warthog logger.go:42: 07:47:21 | examples-with-cassandra/0-install | starting test step 0-install logger.go:42: 07:47:21 | examples-with-cassandra/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:47:21 | examples-with-cassandra/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:47:22 | examples-with-cassandra/0-install | >>>> Creating namespace kuttl-test-moved-warthog logger.go:42: 07:47:22 | examples-with-cassandra/0-install | kubectl create namespace kuttl-test-moved-warthog 2>&1 | grep -v "already exists" || true logger.go:42: 07:47:22 | examples-with-cassandra/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-moved-warthog 2>&1 | grep -v "already exists" || true logger.go:42: 07:47:22 | examples-with-cassandra/0-install | service/cassandra created logger.go:42: 07:47:22 | examples-with-cassandra/0-install | statefulset.apps/cassandra created logger.go:42: 07:47:22 | examples-with-cassandra/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:47:25 | examples-with-cassandra/0-install | test step completed 0-install logger.go:42: 07:47:25 | examples-with-cassandra/1-install | starting test step 1-install logger.go:42: 07:47:25 | examples-with-cassandra/1-install | Jaeger:kuttl-test-moved-warthog/with-cassandra created logger.go:42: 07:47:42 | examples-with-cassandra/1-install | test step completed 1-install logger.go:42: 07:47:42 | examples-with-cassandra/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:47:42 | examples-with-cassandra/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-cassandra /dev/null] logger.go:42: 07:47:44 | examples-with-cassandra/2-smoke-test | Warning: resource jaegers/with-cassandra is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:47:50 | examples-with-cassandra/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:47:51 | examples-with-cassandra/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:47:51 | examples-with-cassandra/2-smoke-test | job.batch/report-span created logger.go:42: 07:47:51 | examples-with-cassandra/2-smoke-test | job.batch/check-span created logger.go:42: 07:48:03 | examples-with-cassandra/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:48:03 | examples-with-cassandra | examples-with-cassandra events from ns kuttl-test-moved-warthog: logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:22 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-moved-warthog/cassandra-0 to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:22 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.129.2.36/23] from ovn-kubernetes logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:22 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:22 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:22 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:22 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:23 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-moved-warthog/cassandra-1 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:23 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:24 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.128.2.52/23] from ovn-kubernetes logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:24 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:24 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:24 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:28 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-9cvjx Binding Scheduled Successfully assigned kuttl-test-moved-warthog/with-cassandra-cassandra-schema-job-9cvjx to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:28 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job SuccessfulCreate Created pod: with-cassandra-cassandra-schema-job-9cvjx job-controller logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:29 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-9cvjx AddedInterface Add eth0 [10.131.0.37/23] from ovn-kubernetes logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:29 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-9cvjx.spec.containers{with-cassandra-cassandra-schema-job} Pulling Pulling image "jaegertracing/jaeger-cassandra-schema:1.51.0" kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:34 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-9cvjx.spec.containers{with-cassandra-cassandra-schema-job} Pulled Successfully pulled image "jaegertracing/jaeger-cassandra-schema:1.51.0" in 4.699s (4.699s including waiting) kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:34 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-9cvjx.spec.containers{with-cassandra-cassandra-schema-job} Created Created container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:34 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-9cvjx.spec.containers{with-cassandra-cassandra-schema-job} Started Started container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:39 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job Completed Job completed job-controller logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:40 +0000 UTC Normal Pod with-cassandra-54f6ddf98d-pwptp Binding Scheduled Successfully assigned kuttl-test-moved-warthog/with-cassandra-54f6ddf98d-pwptp to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:40 +0000 UTC Normal Pod with-cassandra-54f6ddf98d-pwptp AddedInterface Add eth0 [10.129.2.37/23] from ovn-kubernetes logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:40 +0000 UTC Normal Pod with-cassandra-54f6ddf98d-pwptp.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:40 +0000 UTC Normal Pod with-cassandra-54f6ddf98d-pwptp.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:40 +0000 UTC Normal Pod with-cassandra-54f6ddf98d-pwptp.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:40 +0000 UTC Normal Pod with-cassandra-54f6ddf98d-pwptp.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:40 +0000 UTC Normal Pod with-cassandra-54f6ddf98d-pwptp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:40 +0000 UTC Normal Pod with-cassandra-54f6ddf98d-pwptp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:40 +0000 UTC Normal ReplicaSet.apps with-cassandra-54f6ddf98d SuccessfulCreate Created pod: with-cassandra-54f6ddf98d-pwptp replicaset-controller logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:40 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-54f6ddf98d to 1 deployment-controller logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:45 +0000 UTC Normal Pod with-cassandra-54f6ddf98d-pwptp.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:45 +0000 UTC Normal Pod with-cassandra-54f6ddf98d-pwptp.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:45 +0000 UTC Normal ReplicaSet.apps with-cassandra-54f6ddf98d SuccessfulDelete Deleted pod: with-cassandra-54f6ddf98d-pwptp replicaset-controller logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:45 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled down replica set with-cassandra-54f6ddf98d to 0 from 1 deployment-controller logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:46 +0000 UTC Normal Pod with-cassandra-5cb9db8c59-7s55n Binding Scheduled Successfully assigned kuttl-test-moved-warthog/with-cassandra-5cb9db8c59-7s55n to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:46 +0000 UTC Normal ReplicaSet.apps with-cassandra-5cb9db8c59 SuccessfulCreate Created pod: with-cassandra-5cb9db8c59-7s55n replicaset-controller logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:46 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-5cb9db8c59 to 1 deployment-controller logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:47 +0000 UTC Normal Pod with-cassandra-5cb9db8c59-7s55n AddedInterface Add eth0 [10.129.2.38/23] from ovn-kubernetes logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:47 +0000 UTC Normal Pod with-cassandra-5cb9db8c59-7s55n.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:47 +0000 UTC Normal Pod with-cassandra-5cb9db8c59-7s55n.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:47 +0000 UTC Normal Pod with-cassandra-5cb9db8c59-7s55n.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:47 +0000 UTC Normal Pod with-cassandra-5cb9db8c59-7s55n.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:47 +0000 UTC Normal Pod with-cassandra-5cb9db8c59-7s55n.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:47 +0000 UTC Normal Pod with-cassandra-5cb9db8c59-7s55n.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:51 +0000 UTC Normal Pod check-span-6k5jj Binding Scheduled Successfully assigned kuttl-test-moved-warthog/check-span-6k5jj to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:51 +0000 UTC Normal Pod check-span-6k5jj AddedInterface Add eth0 [10.128.2.53/23] from ovn-kubernetes logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:51 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-6k5jj job-controller logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:51 +0000 UTC Normal Pod report-span-8tfzh Binding Scheduled Successfully assigned kuttl-test-moved-warthog/report-span-8tfzh to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:51 +0000 UTC Normal Pod report-span-8tfzh AddedInterface Add eth0 [10.131.0.38/23] from ovn-kubernetes logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:51 +0000 UTC Normal Pod report-span-8tfzh.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:51 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-8tfzh job-controller logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:52 +0000 UTC Normal Pod check-span-6k5jj.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:52 +0000 UTC Normal Pod check-span-6k5jj.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:52 +0000 UTC Normal Pod check-span-6k5jj.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:52 +0000 UTC Normal Pod report-span-8tfzh.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:47:52 +0000 UTC Normal Pod report-span-8tfzh.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:48:03 | examples-with-cassandra | 2023-12-25 07:48:02 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:48:03 | examples-with-cassandra | Deleting namespace: kuttl-test-moved-warthog === CONT kuttl/harness/examples-with-badger-and-volume logger.go:42: 07:48:17 | examples-with-badger-and-volume | Creating namespace: kuttl-test-optimum-crayfish logger.go:42: 07:48:17 | examples-with-badger-and-volume/0-install | starting test step 0-install logger.go:42: 07:48:17 | examples-with-badger-and-volume/0-install | Jaeger:kuttl-test-optimum-crayfish/with-badger-and-volume created logger.go:42: 07:48:23 | examples-with-badger-and-volume/0-install | test step completed 0-install logger.go:42: 07:48:23 | examples-with-badger-and-volume/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:48:23 | examples-with-badger-and-volume/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger-and-volume /dev/null] logger.go:42: 07:48:24 | examples-with-badger-and-volume/1-smoke-test | Warning: resource jaegers/with-badger-and-volume is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:48:31 | examples-with-badger-and-volume/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:48:31 | examples-with-badger-and-volume/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:48:31 | examples-with-badger-and-volume/1-smoke-test | job.batch/report-span created logger.go:42: 07:48:31 | examples-with-badger-and-volume/1-smoke-test | job.batch/check-span created logger.go:42: 07:58:32 | examples-with-badger-and-volume/1-smoke-test | test step failed 1-smoke-test case.go:364: failed in step 1-smoke-test case.go:366: --- Job:kuttl-test-optimum-crayfish/check-span +++ Job:kuttl-test-optimum-crayfish/check-span @@ -1,8 +1,141 @@ apiVersion: batch/v1 kind: Job metadata: + annotations: + kubectl.kubernetes.io/last-applied-configuration: | + {"apiVersion":"batch/v1","kind":"Job","metadata":{"annotations":{},"name":"check-span","namespace":"kuttl-test-optimum-crayfish"},"spec":{"backoffLimit":15,"template":{"spec":{"containers":[{"command":["./query"],"env":[{"name":"SERVICE_NAME","value":"smoke-test-service"},{"name":"QUERY_HOST","value":"https://with-badger-and-volume-query:443"},{"name":"SECRET_PATH","value":"/var/run/secrets/api-token/token"}],"image":"registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7","name":"asserts-container","volumeMounts":[{"mountPath":"/var/run/secrets/api-token","name":"token-api-volume"}]}],"restartPolicy":"OnFailure","volumes":[{"name":"token-api-volume","secret":{"secretName":"e2e-test"}}]}}}} + labels: + batch.kubernetes.io/controller-uid: ddc25355-5fce-4c23-8847-3e2a20d0c480 + batch.kubernetes.io/job-name: check-span + controller-uid: ddc25355-5fce-4c23-8847-3e2a20d0c480 + job-name: check-span + managedFields: + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:kubectl.kubernetes.io/last-applied-configuration: {} + f:spec: + f:backoffLimit: {} + f:completionMode: {} + f:completions: {} + f:parallelism: {} + f:suspend: {} + f:template: + f:spec: + f:containers: + k:{"name":"asserts-container"}: + .: {} + f:command: {} + f:env: + .: {} + k:{"name":"QUERY_HOST"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SECRET_PATH"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SERVICE_NAME"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:name: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/var/run/secrets/api-token"}: + .: {} + f:mountPath: {} + f:name: {} + f:dnsPolicy: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"token-api-volume"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + manager: kubectl-client-side-apply + operation: Update + time: "2023-12-25T07:48:31Z" + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:status: + f:active: {} + f:ready: {} + f:startTime: {} + f:uncountedTerminatedPods: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2023-12-25T07:57:14Z" name: check-span namespace: kuttl-test-optimum-crayfish +spec: + backoffLimit: 15 + completionMode: NonIndexed + completions: 1 + parallelism: 1 + selector: + matchLabels: + batch.kubernetes.io/controller-uid: ddc25355-5fce-4c23-8847-3e2a20d0c480 + suspend: false + template: + metadata: + creationTimestamp: null + labels: + batch.kubernetes.io/controller-uid: ddc25355-5fce-4c23-8847-3e2a20d0c480 + batch.kubernetes.io/job-name: check-span + controller-uid: ddc25355-5fce-4c23-8847-3e2a20d0c480 + job-name: check-span + spec: + containers: + - command: + - ./query + env: + - name: SERVICE_NAME + value: smoke-test-service + - name: QUERY_HOST + value: https://with-badger-and-volume-query:443 + - name: SECRET_PATH + value: /var/run/secrets/api-token/token + image: registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 + imagePullPolicy: IfNotPresent + name: asserts-container + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /var/run/secrets/api-token + name: token-api-volume + dnsPolicy: ClusterFirst + restartPolicy: OnFailure + schedulerName: default-scheduler + securityContext: {} + terminationGracePeriodSeconds: 30 + volumes: + - name: token-api-volume + secret: + defaultMode: 420 + secretName: e2e-test status: - succeeded: 1 + active: 1 + ready: 1 + startTime: "2023-12-25T07:48:31Z" + uncountedTerminatedPods: {} case.go:366: resource Job:kuttl-test-optimum-crayfish/check-span: .status.succeeded: key is missing from map logger.go:42: 07:58:32 | examples-with-badger-and-volume | examples-with-badger-and-volume events from ns kuttl-test-optimum-crayfish: logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:20 +0000 UTC Normal Pod with-badger-and-volume-6cf6f8d55f-7nxdd Binding Scheduled Successfully assigned kuttl-test-optimum-crayfish/with-badger-and-volume-6cf6f8d55f-7nxdd to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:20 +0000 UTC Normal ReplicaSet.apps with-badger-and-volume-6cf6f8d55f SuccessfulCreate Created pod: with-badger-and-volume-6cf6f8d55f-7nxdd replicaset-controller logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:20 +0000 UTC Normal Deployment.apps with-badger-and-volume ScalingReplicaSet Scaled up replica set with-badger-and-volume-6cf6f8d55f to 1 deployment-controller logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:21 +0000 UTC Normal Pod with-badger-and-volume-6cf6f8d55f-7nxdd AddedInterface Add eth0 [10.129.2.39/23] from ovn-kubernetes logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:21 +0000 UTC Normal Pod with-badger-and-volume-6cf6f8d55f-7nxdd.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:21 +0000 UTC Normal Pod with-badger-and-volume-6cf6f8d55f-7nxdd.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:21 +0000 UTC Normal Pod with-badger-and-volume-6cf6f8d55f-7nxdd.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:21 +0000 UTC Normal Pod with-badger-and-volume-6cf6f8d55f-7nxdd.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:21 +0000 UTC Normal Pod with-badger-and-volume-6cf6f8d55f-7nxdd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:21 +0000 UTC Normal Pod with-badger-and-volume-6cf6f8d55f-7nxdd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:31 +0000 UTC Normal Pod check-span-n68l5 Binding Scheduled Successfully assigned kuttl-test-optimum-crayfish/check-span-n68l5 to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:31 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-n68l5 job-controller logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:31 +0000 UTC Normal Pod report-span-hqcw9 Binding Scheduled Successfully assigned kuttl-test-optimum-crayfish/report-span-hqcw9 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:31 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-hqcw9 job-controller logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:32 +0000 UTC Normal Pod check-span-n68l5 AddedInterface Add eth0 [10.131.0.39/23] from ovn-kubernetes logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:32 +0000 UTC Normal Pod check-span-n68l5.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:32 +0000 UTC Normal Pod check-span-n68l5.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:32 +0000 UTC Normal Pod check-span-n68l5.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:32 +0000 UTC Normal Pod report-span-hqcw9 AddedInterface Add eth0 [10.128.2.54/23] from ovn-kubernetes logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:32 +0000 UTC Normal Pod report-span-hqcw9.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:32 +0000 UTC Normal Pod report-span-hqcw9.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:48:32 +0000 UTC Normal Pod report-span-hqcw9.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:58:32 | examples-with-badger-and-volume | 2023-12-25 07:53:53 +0000 UTC Warning Pod check-span-n68l5.spec.containers{asserts-container} BackOff Back-off restarting failed container asserts-container in pod check-span-n68l5_kuttl-test-optimum-crayfish(3317ebf4-7efd-4a67-8124-b21469ddac9a) kubelet logger.go:42: 07:58:32 | examples-with-badger-and-volume | Deleting namespace: kuttl-test-optimum-crayfish === CONT kuttl/harness/examples-with-badger logger.go:42: 07:58:39 | examples-with-badger | Creating namespace: kuttl-test-quiet-jackass logger.go:42: 07:58:39 | examples-with-badger/0-install | starting test step 0-install logger.go:42: 07:58:39 | examples-with-badger/0-install | Jaeger:kuttl-test-quiet-jackass/with-badger created logger.go:42: 07:58:44 | examples-with-badger/0-install | test step completed 0-install logger.go:42: 07:58:44 | examples-with-badger/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:58:44 | examples-with-badger/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger /dev/null] logger.go:42: 07:58:46 | examples-with-badger/1-smoke-test | Warning: resource jaegers/with-badger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:58:52 | examples-with-badger/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:58:52 | examples-with-badger/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:58:53 | examples-with-badger/1-smoke-test | job.batch/report-span created logger.go:42: 07:58:53 | examples-with-badger/1-smoke-test | job.batch/check-span created logger.go:42: 07:59:05 | examples-with-badger/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:59:05 | examples-with-badger | examples-with-badger events from ns kuttl-test-quiet-jackass: logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:43 +0000 UTC Normal Pod with-badger-b85c4b89f-xkl92 Binding Scheduled Successfully assigned kuttl-test-quiet-jackass/with-badger-b85c4b89f-xkl92 to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:43 +0000 UTC Normal Pod with-badger-b85c4b89f-xkl92 AddedInterface Add eth0 [10.129.2.40/23] from ovn-kubernetes logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:43 +0000 UTC Normal Pod with-badger-b85c4b89f-xkl92.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:43 +0000 UTC Normal Pod with-badger-b85c4b89f-xkl92.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:43 +0000 UTC Normal Pod with-badger-b85c4b89f-xkl92.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:43 +0000 UTC Normal Pod with-badger-b85c4b89f-xkl92.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:43 +0000 UTC Normal Pod with-badger-b85c4b89f-xkl92.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:43 +0000 UTC Normal Pod with-badger-b85c4b89f-xkl92.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:43 +0000 UTC Normal ReplicaSet.apps with-badger-b85c4b89f SuccessfulCreate Created pod: with-badger-b85c4b89f-xkl92 replicaset-controller logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:43 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-b85c4b89f to 1 deployment-controller logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:48 +0000 UTC Normal Pod with-badger-b85c4b89f-xkl92.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:48 +0000 UTC Normal Pod with-badger-b85c4b89f-xkl92.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:48 +0000 UTC Normal ReplicaSet.apps with-badger-b85c4b89f SuccessfulDelete Deleted pod: with-badger-b85c4b89f-xkl92 replicaset-controller logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:48 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled down replica set with-badger-b85c4b89f to 0 from 1 deployment-controller logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:49 +0000 UTC Normal Pod with-badger-6cbd9df56f-pf7kk Binding Scheduled Successfully assigned kuttl-test-quiet-jackass/with-badger-6cbd9df56f-pf7kk to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:49 +0000 UTC Normal Pod with-badger-6cbd9df56f-pf7kk AddedInterface Add eth0 [10.129.2.41/23] from ovn-kubernetes logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:49 +0000 UTC Normal Pod with-badger-6cbd9df56f-pf7kk.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:49 +0000 UTC Normal Pod with-badger-6cbd9df56f-pf7kk.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:49 +0000 UTC Normal ReplicaSet.apps with-badger-6cbd9df56f SuccessfulCreate Created pod: with-badger-6cbd9df56f-pf7kk replicaset-controller logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:49 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-6cbd9df56f to 1 deployment-controller logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:50 +0000 UTC Normal Pod with-badger-6cbd9df56f-pf7kk.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:50 +0000 UTC Normal Pod with-badger-6cbd9df56f-pf7kk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:50 +0000 UTC Normal Pod with-badger-6cbd9df56f-pf7kk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:50 +0000 UTC Normal Pod with-badger-6cbd9df56f-pf7kk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:53 +0000 UTC Normal Pod check-span-76gjg Binding Scheduled Successfully assigned kuttl-test-quiet-jackass/check-span-76gjg to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:53 +0000 UTC Normal Pod check-span-76gjg AddedInterface Add eth0 [10.131.0.40/23] from ovn-kubernetes logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:53 +0000 UTC Normal Pod check-span-76gjg.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:53 +0000 UTC Normal Pod check-span-76gjg.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:53 +0000 UTC Normal Pod check-span-76gjg.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:53 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-76gjg job-controller logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:53 +0000 UTC Normal Pod report-span-btfvw Binding Scheduled Successfully assigned kuttl-test-quiet-jackass/report-span-btfvw to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:53 +0000 UTC Normal Pod report-span-btfvw AddedInterface Add eth0 [10.128.2.55/23] from ovn-kubernetes logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:53 +0000 UTC Normal Pod report-span-btfvw.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:53 +0000 UTC Normal Pod report-span-btfvw.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:53 +0000 UTC Normal Pod report-span-btfvw.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:58:53 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-btfvw job-controller logger.go:42: 07:59:05 | examples-with-badger | 2023-12-25 07:59:04 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:59:05 | examples-with-badger | Deleting namespace: kuttl-test-quiet-jackass === CONT kuttl/harness/examples-simplest logger.go:42: 07:59:18 | examples-simplest | Creating namespace: kuttl-test-game-blowfish logger.go:42: 07:59:18 | examples-simplest/0-install | starting test step 0-install logger.go:42: 07:59:18 | examples-simplest/0-install | Jaeger:kuttl-test-game-blowfish/simplest created logger.go:42: 07:59:23 | examples-simplest/0-install | test step completed 0-install logger.go:42: 07:59:23 | examples-simplest/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:59:23 | examples-simplest/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:59:25 | examples-simplest/1-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:59:31 | examples-simplest/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:59:32 | examples-simplest/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:59:32 | examples-simplest/1-smoke-test | job.batch/report-span created logger.go:42: 07:59:32 | examples-simplest/1-smoke-test | job.batch/check-span created logger.go:42: 07:59:44 | examples-simplest/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:59:44 | examples-simplest | examples-simplest events from ns kuttl-test-game-blowfish: logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:22 +0000 UTC Normal Pod simplest-59586c98c5-wkkw7 Binding Scheduled Successfully assigned kuttl-test-game-blowfish/simplest-59586c98c5-wkkw7 to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:22 +0000 UTC Normal Pod simplest-59586c98c5-wkkw7 AddedInterface Add eth0 [10.129.2.42/23] from ovn-kubernetes logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:22 +0000 UTC Normal Pod simplest-59586c98c5-wkkw7.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:22 +0000 UTC Normal Pod simplest-59586c98c5-wkkw7.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:22 +0000 UTC Normal Pod simplest-59586c98c5-wkkw7.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:22 +0000 UTC Normal Pod simplest-59586c98c5-wkkw7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:22 +0000 UTC Normal Pod simplest-59586c98c5-wkkw7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:22 +0000 UTC Normal Pod simplest-59586c98c5-wkkw7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:22 +0000 UTC Normal ReplicaSet.apps simplest-59586c98c5 SuccessfulCreate Created pod: simplest-59586c98c5-wkkw7 replicaset-controller logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:22 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-59586c98c5 to 1 deployment-controller logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:26 +0000 UTC Normal Pod simplest-59586c98c5-wkkw7.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:26 +0000 UTC Normal Pod simplest-59586c98c5-wkkw7.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:26 +0000 UTC Normal ReplicaSet.apps simplest-59586c98c5 SuccessfulDelete Deleted pod: simplest-59586c98c5-wkkw7 replicaset-controller logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:26 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-59586c98c5 to 0 from 1 deployment-controller logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:27 +0000 UTC Normal Pod simplest-67fb64bfbf-xsjpd Binding Scheduled Successfully assigned kuttl-test-game-blowfish/simplest-67fb64bfbf-xsjpd to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:27 +0000 UTC Normal ReplicaSet.apps simplest-67fb64bfbf SuccessfulCreate Created pod: simplest-67fb64bfbf-xsjpd replicaset-controller logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:27 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-67fb64bfbf to 1 deployment-controller logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:28 +0000 UTC Normal Pod simplest-67fb64bfbf-xsjpd AddedInterface Add eth0 [10.129.2.43/23] from ovn-kubernetes logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:28 +0000 UTC Normal Pod simplest-67fb64bfbf-xsjpd.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:28 +0000 UTC Normal Pod simplest-67fb64bfbf-xsjpd.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:28 +0000 UTC Normal Pod simplest-67fb64bfbf-xsjpd.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:28 +0000 UTC Normal Pod simplest-67fb64bfbf-xsjpd.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:28 +0000 UTC Normal Pod simplest-67fb64bfbf-xsjpd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:28 +0000 UTC Normal Pod simplest-67fb64bfbf-xsjpd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:32 +0000 UTC Normal Pod check-span-6r24n Binding Scheduled Successfully assigned kuttl-test-game-blowfish/check-span-6r24n to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:32 +0000 UTC Normal Pod check-span-6r24n AddedInterface Add eth0 [10.131.0.41/23] from ovn-kubernetes logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:32 +0000 UTC Normal Pod check-span-6r24n.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:32 +0000 UTC Normal Pod check-span-6r24n.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:32 +0000 UTC Normal Pod check-span-6r24n.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:32 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-6r24n job-controller logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:32 +0000 UTC Normal Pod report-span-hmb85 Binding Scheduled Successfully assigned kuttl-test-game-blowfish/report-span-hmb85 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:32 +0000 UTC Normal Pod report-span-hmb85 AddedInterface Add eth0 [10.128.2.56/23] from ovn-kubernetes logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:32 +0000 UTC Normal Pod report-span-hmb85.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:32 +0000 UTC Normal Pod report-span-hmb85.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:32 +0000 UTC Normal Pod report-span-hmb85.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:32 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-hmb85 job-controller logger.go:42: 07:59:44 | examples-simplest | 2023-12-25 07:59:43 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:59:44 | examples-simplest | Deleting namespace: kuttl-test-game-blowfish === CONT kuttl/harness/examples-simple-prod-with-volumes logger.go:42: 07:59:57 | examples-simple-prod-with-volumes | Ignoring 03-check-volume.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:59:57 | examples-simple-prod-with-volumes | Creating namespace: kuttl-test-subtle-ocelot logger.go:42: 07:59:57 | examples-simple-prod-with-volumes/1-install | starting test step 1-install logger.go:42: 07:59:57 | examples-simple-prod-with-volumes/1-install | Jaeger:kuttl-test-subtle-ocelot/simple-prod created logger.go:42: 08:00:34 | examples-simple-prod-with-volumes/1-install | test step completed 1-install logger.go:42: 08:00:34 | examples-simple-prod-with-volumes/2-smoke-test | starting test step 2-smoke-test logger.go:42: 08:00:34 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 08:00:36 | examples-simple-prod-with-volumes/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:00:42 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 08:00:43 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:00:43 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/report-span created logger.go:42: 08:00:43 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/check-span created logger.go:42: 08:00:54 | examples-simple-prod-with-volumes/2-smoke-test | test step completed 2-smoke-test logger.go:42: 08:00:54 | examples-simple-prod-with-volumes/3-check-volume | starting test step 3-check-volume logger.go:42: 08:00:54 | examples-simple-prod-with-volumes/3-check-volume | running command: [sh -c kubectl exec $(kubectl get pods -n $NAMESPACE -l app=jaeger -l app.kubernetes.io/component=collector -o yaml | /tmp/jaeger-tests/bin/yq e '.items[0].metadata.name') -n $NAMESPACE -- ls /usr/share/elasticsearch/data] logger.go:42: 08:00:55 | examples-simple-prod-with-volumes/3-check-volume | test step completed 3-check-volume logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | examples-simple-prod-with-volumes events from ns kuttl-test-subtle-ocelot: logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:03 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestsubtleocelotsimpleprod-1-5f85b479ff SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestsubtleocelotsimpleprod-1-5f85b4xrm7c replicaset-controller logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsubtleocelotsimpleprod-1-5f85b4xrm7c Binding Scheduled Successfully assigned kuttl-test-subtle-ocelot/elasticsearch-cdm-kuttltestsubtleocelotsimpleprod-1-5f85b4xrm7c to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsubtleocelotsimpleprod-1-5f85b4xrm7c AddedInterface Add eth0 [10.129.2.45/23] from ovn-kubernetes logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsubtleocelotsimpleprod-1-5f85b4xrm7c.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:03 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestsubtleocelotsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestsubtleocelotsimpleprod-1-5f85b479ff to 1 deployment-controller logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsubtleocelotsimpleprod-1-5f85b4xrm7c.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsubtleocelotsimpleprod-1-5f85b4xrm7c.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsubtleocelotsimpleprod-1-5f85b4xrm7c.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsubtleocelotsimpleprod-1-5f85b4xrm7c.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsubtleocelotsimpleprod-1-5f85b4xrm7c.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:19 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsubtleocelotsimpleprod-1-5f85b4xrm7c.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:30 +0000 UTC Normal Pod simple-prod-collector-56887d4c49-nzcjm Binding Scheduled Successfully assigned kuttl-test-subtle-ocelot/simple-prod-collector-56887d4c49-nzcjm to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:30 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-56887d4c49 SuccessfulCreate Created pod: simple-prod-collector-56887d4c49-nzcjm replicaset-controller logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:30 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-56887d4c49 to 1 deployment-controller logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:30 +0000 UTC Normal Pod simple-prod-query-69fbcbf5df-fv8vs Binding Scheduled Successfully assigned kuttl-test-subtle-ocelot/simple-prod-query-69fbcbf5df-fv8vs to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:30 +0000 UTC Normal ReplicaSet.apps simple-prod-query-69fbcbf5df SuccessfulCreate Created pod: simple-prod-query-69fbcbf5df-fv8vs replicaset-controller logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:30 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-69fbcbf5df to 1 deployment-controller logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:31 +0000 UTC Warning Pod simple-prod-collector-56887d4c49-nzcjm FailedMount MountVolume.SetUp failed for volume "certs" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:31 +0000 UTC Warning Pod simple-prod-query-69fbcbf5df-fv8vs FailedMount MountVolume.SetUp failed for volume "simple-prod-service-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:31 +0000 UTC Warning Pod simple-prod-query-69fbcbf5df-fv8vs FailedMount MountVolume.SetUp failed for volume "simple-prod-trusted-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:32 +0000 UTC Normal Pod simple-prod-collector-56887d4c49-nzcjm AddedInterface Add eth0 [10.128.2.57/23] from ovn-kubernetes logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:32 +0000 UTC Normal Pod simple-prod-collector-56887d4c49-nzcjm.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:32 +0000 UTC Normal Pod simple-prod-collector-56887d4c49-nzcjm.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:32 +0000 UTC Normal Pod simple-prod-collector-56887d4c49-nzcjm.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:32 +0000 UTC Normal Pod simple-prod-query-69fbcbf5df-fv8vs AddedInterface Add eth0 [10.131.0.42/23] from ovn-kubernetes logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:32 +0000 UTC Normal Pod simple-prod-query-69fbcbf5df-fv8vs.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:32 +0000 UTC Normal Pod simple-prod-query-69fbcbf5df-fv8vs.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:32 +0000 UTC Normal Pod simple-prod-query-69fbcbf5df-fv8vs.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:32 +0000 UTC Normal Pod simple-prod-query-69fbcbf5df-fv8vs.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:32 +0000 UTC Normal Pod simple-prod-query-69fbcbf5df-fv8vs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:32 +0000 UTC Normal Pod simple-prod-query-69fbcbf5df-fv8vs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:32 +0000 UTC Normal Pod simple-prod-query-69fbcbf5df-fv8vs.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:32 +0000 UTC Normal Pod simple-prod-query-69fbcbf5df-fv8vs.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:32 +0000 UTC Normal Pod simple-prod-query-69fbcbf5df-fv8vs.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:37 +0000 UTC Normal Pod simple-prod-query-69fbcbf5df-fv8vs.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:37 +0000 UTC Normal Pod simple-prod-query-69fbcbf5df-fv8vs.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:37 +0000 UTC Normal Pod simple-prod-query-69fbcbf5df-fv8vs.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:37 +0000 UTC Normal ReplicaSet.apps simple-prod-query-69fbcbf5df SuccessfulDelete Deleted pod: simple-prod-query-69fbcbf5df-fv8vs replicaset-controller logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:37 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-69fbcbf5df to 0 from 1 deployment-controller logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:38 +0000 UTC Normal Pod simple-prod-query-8657fb755f-2m45p Binding Scheduled Successfully assigned kuttl-test-subtle-ocelot/simple-prod-query-8657fb755f-2m45p to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:38 +0000 UTC Normal Pod simple-prod-query-8657fb755f-2m45p AddedInterface Add eth0 [10.131.0.43/23] from ovn-kubernetes logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:38 +0000 UTC Normal Pod simple-prod-query-8657fb755f-2m45p.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:38 +0000 UTC Normal Pod simple-prod-query-8657fb755f-2m45p.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:38 +0000 UTC Normal Pod simple-prod-query-8657fb755f-2m45p.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:38 +0000 UTC Normal Pod simple-prod-query-8657fb755f-2m45p.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:38 +0000 UTC Normal Pod simple-prod-query-8657fb755f-2m45p.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:38 +0000 UTC Normal Pod simple-prod-query-8657fb755f-2m45p.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:38 +0000 UTC Normal ReplicaSet.apps simple-prod-query-8657fb755f SuccessfulCreate Created pod: simple-prod-query-8657fb755f-2m45p replicaset-controller logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:38 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-8657fb755f to 1 deployment-controller logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:39 +0000 UTC Normal Pod simple-prod-query-8657fb755f-2m45p.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:39 +0000 UTC Normal Pod simple-prod-query-8657fb755f-2m45p.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:39 +0000 UTC Normal Pod simple-prod-query-8657fb755f-2m45p.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:43 +0000 UTC Normal Pod check-span-zsqjd Binding Scheduled Successfully assigned kuttl-test-subtle-ocelot/check-span-zsqjd to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:43 +0000 UTC Normal Pod check-span-zsqjd AddedInterface Add eth0 [10.128.2.59/23] from ovn-kubernetes logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:43 +0000 UTC Normal Pod check-span-zsqjd.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:43 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-zsqjd job-controller logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:43 +0000 UTC Normal Pod report-span-tjwtr Binding Scheduled Successfully assigned kuttl-test-subtle-ocelot/report-span-tjwtr to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:43 +0000 UTC Normal Pod report-span-tjwtr AddedInterface Add eth0 [10.128.2.58/23] from ovn-kubernetes logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:43 +0000 UTC Normal Pod report-span-tjwtr.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:43 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-tjwtr job-controller logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:44 +0000 UTC Normal Pod check-span-zsqjd.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:44 +0000 UTC Normal Pod check-span-zsqjd.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:44 +0000 UTC Normal Pod report-span-tjwtr.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:44 +0000 UTC Normal Pod report-span-tjwtr.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | 2023-12-25 08:00:54 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:00:55 | examples-simple-prod-with-volumes | Deleting namespace: kuttl-test-subtle-ocelot === CONT kuttl/harness/examples-business-application-injected-sidecar logger.go:42: 08:01:08 | examples-business-application-injected-sidecar | Creating namespace: kuttl-test-fancy-ape logger.go:42: 08:01:08 | examples-business-application-injected-sidecar/0-install | starting test step 0-install logger.go:42: 08:01:08 | examples-business-application-injected-sidecar/0-install | Deployment:kuttl-test-fancy-ape/myapp created logger.go:42: 08:01:08 | examples-business-application-injected-sidecar/0-install | test step completed 0-install logger.go:42: 08:01:08 | examples-business-application-injected-sidecar/1-install | starting test step 1-install logger.go:42: 08:01:08 | examples-business-application-injected-sidecar/1-install | Jaeger:kuttl-test-fancy-ape/simplest created logger.go:42: 08:01:18 | examples-business-application-injected-sidecar/1-install | test step completed 1-install logger.go:42: 08:01:18 | examples-business-application-injected-sidecar/2-smoke-test | starting test step 2-smoke-test logger.go:42: 08:01:18 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 08:01:20 | examples-business-application-injected-sidecar/2-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:01:26 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 08:01:27 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:01:27 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/report-span created logger.go:42: 08:01:27 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/check-span created logger.go:42: 08:01:33 | examples-business-application-injected-sidecar/2-smoke-test | test step completed 2-smoke-test logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | examples-business-application-injected-sidecar events from ns kuttl-test-fancy-ape: logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:08 +0000 UTC Normal Pod myapp-679f79d5f8-vfwtn Binding Scheduled Successfully assigned kuttl-test-fancy-ape/myapp-679f79d5f8-vfwtn to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:08 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulCreate Created pod: myapp-679f79d5f8-vfwtn replicaset-controller logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:08 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-679f79d5f8 to 1 deployment-controller logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:09 +0000 UTC Normal Pod myapp-679f79d5f8-vfwtn AddedInterface Add eth0 [10.129.2.46/23] from ovn-kubernetes logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:09 +0000 UTC Normal Pod myapp-679f79d5f8-vfwtn.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:09 +0000 UTC Normal Pod myapp-76786b8c79-jm7qx Binding Scheduled Successfully assigned kuttl-test-fancy-ape/myapp-76786b8c79-jm7qx to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:09 +0000 UTC Normal ReplicaSet.apps myapp-76786b8c79 SuccessfulCreate Created pod: myapp-76786b8c79-jm7qx replicaset-controller logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:09 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-76786b8c79 to 1 deployment-controller logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:10 +0000 UTC Warning Pod myapp-76786b8c79-jm7qx FailedMount MountVolume.SetUp failed for volume "simplest-trusted-ca" : configmap "simplest-trusted-ca" not found kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:10 +0000 UTC Warning Pod myapp-76786b8c79-jm7qx FailedMount MountVolume.SetUp failed for volume "simplest-service-ca" : configmap "simplest-service-ca" not found kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:13 +0000 UTC Normal Pod myapp-679f79d5f8-vfwtn.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 4.175s (4.175s including waiting) kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:13 +0000 UTC Normal Pod myapp-679f79d5f8-vfwtn.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:13 +0000 UTC Normal Pod myapp-679f79d5f8-vfwtn.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:17 +0000 UTC Normal Pod myapp-76786b8c79-jm7qx AddedInterface Add eth0 [10.131.0.44/23] from ovn-kubernetes logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:17 +0000 UTC Normal Pod myapp-76786b8c79-jm7qx.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:17 +0000 UTC Normal Pod simplest-6d5f8c8f5-xfxrp Binding Scheduled Successfully assigned kuttl-test-fancy-ape/simplest-6d5f8c8f5-xfxrp to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:17 +0000 UTC Normal Pod simplest-6d5f8c8f5-xfxrp AddedInterface Add eth0 [10.129.2.47/23] from ovn-kubernetes logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:17 +0000 UTC Normal Pod simplest-6d5f8c8f5-xfxrp.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:17 +0000 UTC Normal Pod simplest-6d5f8c8f5-xfxrp.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:17 +0000 UTC Normal Pod simplest-6d5f8c8f5-xfxrp.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:17 +0000 UTC Normal Pod simplest-6d5f8c8f5-xfxrp.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:17 +0000 UTC Normal Pod simplest-6d5f8c8f5-xfxrp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:17 +0000 UTC Normal Pod simplest-6d5f8c8f5-xfxrp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:17 +0000 UTC Normal ReplicaSet.apps simplest-6d5f8c8f5 SuccessfulCreate Created pod: simplest-6d5f8c8f5-xfxrp replicaset-controller logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:17 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-6d5f8c8f5 to 1 deployment-controller logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:20 +0000 UTC Warning Pod myapp-679f79d5f8-vfwtn.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.129.2.46:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:21 +0000 UTC Normal Pod myapp-76786b8c79-jm7qx.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.645s (3.645s including waiting) kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:21 +0000 UTC Normal Pod myapp-76786b8c79-jm7qx.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:21 +0000 UTC Normal Pod myapp-76786b8c79-jm7qx.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:21 +0000 UTC Normal Pod myapp-76786b8c79-jm7qx.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:21 +0000 UTC Normal Pod myapp-76786b8c79-jm7qx.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:21 +0000 UTC Normal Pod myapp-76786b8c79-jm7qx.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:21 +0000 UTC Normal Pod simplest-6d5f8c8f5-xfxrp.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:21 +0000 UTC Normal Pod simplest-6d5f8c8f5-xfxrp.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:21 +0000 UTC Normal ReplicaSet.apps simplest-6d5f8c8f5 SuccessfulDelete Deleted pod: simplest-6d5f8c8f5-xfxrp replicaset-controller logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:21 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-6d5f8c8f5 to 0 from 1 deployment-controller logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:22 +0000 UTC Normal Pod myapp-679f79d5f8-vfwtn.spec.containers{myapp} Killing Stopping container myapp kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:22 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulDelete Deleted pod: myapp-679f79d5f8-vfwtn replicaset-controller logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:22 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled down replica set myapp-679f79d5f8 to 0 from 1 deployment-controller logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:22 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-9b45495cd to 1 deployment-controller logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:23 +0000 UTC Normal Pod simplest-9b45495cd-gx8lx Binding Scheduled Successfully assigned kuttl-test-fancy-ape/simplest-9b45495cd-gx8lx to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:23 +0000 UTC Normal Pod simplest-9b45495cd-gx8lx AddedInterface Add eth0 [10.129.2.48/23] from ovn-kubernetes logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:23 +0000 UTC Normal Pod simplest-9b45495cd-gx8lx.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:23 +0000 UTC Normal Pod simplest-9b45495cd-gx8lx.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:23 +0000 UTC Normal Pod simplest-9b45495cd-gx8lx.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:23 +0000 UTC Normal Pod simplest-9b45495cd-gx8lx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:23 +0000 UTC Normal Pod simplest-9b45495cd-gx8lx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:23 +0000 UTC Normal Pod simplest-9b45495cd-gx8lx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:23 +0000 UTC Normal ReplicaSet.apps simplest-9b45495cd SuccessfulCreate Created pod: simplest-9b45495cd-gx8lx replicaset-controller logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:27 +0000 UTC Normal Pod check-span-c9w8r Binding Scheduled Successfully assigned kuttl-test-fancy-ape/check-span-c9w8r to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:27 +0000 UTC Normal Pod check-span-c9w8r AddedInterface Add eth0 [10.129.2.49/23] from ovn-kubernetes logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:27 +0000 UTC Normal Pod check-span-c9w8r.spec.containers{asserts-container} Pulling Pulling image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:27 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-c9w8r job-controller logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:27 +0000 UTC Normal Pod report-span-6ff69 Binding Scheduled Successfully assigned kuttl-test-fancy-ape/report-span-6ff69 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:27 +0000 UTC Normal Pod report-span-6ff69 AddedInterface Add eth0 [10.128.2.60/23] from ovn-kubernetes logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:27 +0000 UTC Normal Pod report-span-6ff69.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:27 +0000 UTC Normal Pod report-span-6ff69.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:27 +0000 UTC Normal Pod report-span-6ff69.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:27 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-6ff69 job-controller logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:28 +0000 UTC Warning Pod myapp-76786b8c79-jm7qx.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.131.0.44:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:30 +0000 UTC Normal Pod check-span-c9w8r.spec.containers{asserts-container} Pulled Successfully pulled image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" in 2.189s (2.189s including waiting) kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:30 +0000 UTC Normal Pod check-span-c9w8r.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:30 +0000 UTC Normal Pod check-span-c9w8r.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | 2023-12-25 08:01:33 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:01:33 | examples-business-application-injected-sidecar | Deleting namespace: kuttl-test-fancy-ape === CONT kuttl/harness/examples-service-types logger.go:42: 08:01:46 | examples-service-types | Creating namespace: kuttl-test-splendid-sheepdog logger.go:42: 08:01:46 | examples-service-types/0-install | starting test step 0-install logger.go:42: 08:01:47 | examples-service-types/0-install | Jaeger:kuttl-test-splendid-sheepdog/service-types created logger.go:42: 08:01:53 | examples-service-types/0-install | test step completed 0-install logger.go:42: 08:01:53 | examples-service-types/1-smoke-test | starting test step 1-smoke-test logger.go:42: 08:01:53 | examples-service-types/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE service-types /dev/null] logger.go:42: 08:01:54 | examples-service-types/1-smoke-test | Warning: resource jaegers/service-types is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:02:00 | examples-service-types/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://service-types-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 08:02:01 | examples-service-types/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:02:01 | examples-service-types/1-smoke-test | job.batch/report-span created logger.go:42: 08:02:01 | examples-service-types/1-smoke-test | job.batch/check-span created logger.go:42: 08:02:14 | examples-service-types/1-smoke-test | test step completed 1-smoke-test logger.go:42: 08:02:14 | examples-service-types/2- | starting test step 2- logger.go:42: 08:02:14 | examples-service-types/2- | test step completed 2- logger.go:42: 08:02:14 | examples-service-types | examples-service-types events from ns kuttl-test-splendid-sheepdog: logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:50 +0000 UTC Normal Pod service-types-6bcb6466bd-b6pd2 Binding Scheduled Successfully assigned kuttl-test-splendid-sheepdog/service-types-6bcb6466bd-b6pd2 to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:50 +0000 UTC Normal ReplicaSet.apps service-types-6bcb6466bd SuccessfulCreate Created pod: service-types-6bcb6466bd-b6pd2 replicaset-controller logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:50 +0000 UTC Normal Service service-types-collector EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:50 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-6bcb6466bd to 1 deployment-controller logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:51 +0000 UTC Normal Pod service-types-6bcb6466bd-b6pd2 AddedInterface Add eth0 [10.129.2.50/23] from ovn-kubernetes logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:51 +0000 UTC Normal Pod service-types-6bcb6466bd-b6pd2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:51 +0000 UTC Normal Pod service-types-6bcb6466bd-b6pd2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:51 +0000 UTC Normal Pod service-types-6bcb6466bd-b6pd2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:51 +0000 UTC Normal Pod service-types-6bcb6466bd-b6pd2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:51 +0000 UTC Normal Pod service-types-6bcb6466bd-b6pd2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:51 +0000 UTC Normal Pod service-types-6bcb6466bd-b6pd2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:53 +0000 UTC Normal Service service-types-collector EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:53 +0000 UTC Normal Service service-types-query EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:56 +0000 UTC Normal Service service-types-query EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:57 +0000 UTC Normal Pod service-types-6bcb6466bd-b6pd2.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:57 +0000 UTC Normal Pod service-types-6bcb6466bd-b6pd2.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:57 +0000 UTC Normal ReplicaSet.apps service-types-6bcb6466bd SuccessfulDelete Deleted pod: service-types-6bcb6466bd-b6pd2 replicaset-controller logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:57 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled down replica set service-types-6bcb6466bd to 0 from 1 deployment-controller logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:58 +0000 UTC Normal Pod service-types-755bb769fb-f9hrk Binding Scheduled Successfully assigned kuttl-test-splendid-sheepdog/service-types-755bb769fb-f9hrk to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:58 +0000 UTC Normal Pod service-types-755bb769fb-f9hrk AddedInterface Add eth0 [10.129.2.51/23] from ovn-kubernetes logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:58 +0000 UTC Normal Pod service-types-755bb769fb-f9hrk.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:58 +0000 UTC Normal Pod service-types-755bb769fb-f9hrk.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:58 +0000 UTC Normal Pod service-types-755bb769fb-f9hrk.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:58 +0000 UTC Normal Pod service-types-755bb769fb-f9hrk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:58 +0000 UTC Normal Pod service-types-755bb769fb-f9hrk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:58 +0000 UTC Normal Pod service-types-755bb769fb-f9hrk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:58 +0000 UTC Normal ReplicaSet.apps service-types-755bb769fb SuccessfulCreate Created pod: service-types-755bb769fb-f9hrk replicaset-controller logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:01:58 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-755bb769fb to 1 deployment-controller logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:02:01 +0000 UTC Normal Pod check-span-4k54g Binding Scheduled Successfully assigned kuttl-test-splendid-sheepdog/check-span-4k54g to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:02:01 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-4k54g job-controller logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:02:01 +0000 UTC Normal Pod report-span-2vn6c Binding Scheduled Successfully assigned kuttl-test-splendid-sheepdog/report-span-2vn6c to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:02:01 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-2vn6c job-controller logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:02:02 +0000 UTC Normal Pod check-span-4k54g AddedInterface Add eth0 [10.131.0.45/23] from ovn-kubernetes logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:02:02 +0000 UTC Normal Pod check-span-4k54g.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:02:02 +0000 UTC Normal Pod check-span-4k54g.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:02:02 +0000 UTC Normal Pod check-span-4k54g.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:02:02 +0000 UTC Normal Pod report-span-2vn6c AddedInterface Add eth0 [10.128.2.61/23] from ovn-kubernetes logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:02:02 +0000 UTC Normal Pod report-span-2vn6c.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:02:02 +0000 UTC Normal Pod report-span-2vn6c.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:02:02 +0000 UTC Normal Pod report-span-2vn6c.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:02:14 | examples-service-types | 2023-12-25 08:02:13 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:02:14 | examples-service-types | Deleting namespace: kuttl-test-splendid-sheepdog === CONT kuttl/harness/examples-openshift-with-htpasswd logger.go:42: 08:02:44 | examples-openshift-with-htpasswd | Ignoring 00-install.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:02:44 | examples-openshift-with-htpasswd | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:02:44 | examples-openshift-with-htpasswd | Creating namespace: kuttl-test-ultimate-lizard logger.go:42: 08:02:44 | examples-openshift-with-htpasswd/0-install | starting test step 0-install logger.go:42: 08:02:44 | examples-openshift-with-htpasswd/0-install | Secret:kuttl-test-ultimate-lizard/htpasswd created logger.go:42: 08:02:44 | examples-openshift-with-htpasswd/0-install | test step completed 0-install logger.go:42: 08:02:44 | examples-openshift-with-htpasswd/1-install | starting test step 1-install logger.go:42: 08:02:44 | examples-openshift-with-htpasswd/1-install | Jaeger:kuttl-test-ultimate-lizard/with-htpasswd created logger.go:42: 08:02:50 | examples-openshift-with-htpasswd/1-install | test step completed 1-install logger.go:42: 08:02:50 | examples-openshift-with-htpasswd/2-check-unsecured | starting test step 2-check-unsecured logger.go:42: 08:02:50 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [./ensure-ingress-host.sh] logger.go:42: 08:02:50 | examples-openshift-with-htpasswd/2-check-unsecured | Checking the Ingress host value was populated logger.go:42: 08:02:50 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 0 logger.go:42: 08:02:50 | examples-openshift-with-htpasswd/2-check-unsecured | Hostname is with-htpasswd-kuttl-test-ultimate-lizard.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com logger.go:42: 08:02:50 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 08:02:50 | examples-openshift-with-htpasswd/2-check-unsecured | Checking an expected HTTP response logger.go:42: 08:02:50 | examples-openshift-with-htpasswd/2-check-unsecured | Running in OpenShift logger.go:42: 08:02:50 | examples-openshift-with-htpasswd/2-check-unsecured | Not using any secret logger.go:42: 08:02:50 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 1/30 the https://with-htpasswd-kuttl-test-ultimate-lizard.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:02:50 | examples-openshift-with-htpasswd/2-check-unsecured | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 08:02:50 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 2/30 the https://with-htpasswd-kuttl-test-ultimate-lizard.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:02:50 | examples-openshift-with-htpasswd/2-check-unsecured | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 08:03:00 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 3/30 the https://with-htpasswd-kuttl-test-ultimate-lizard.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:03:00 | examples-openshift-with-htpasswd/2-check-unsecured | curl response asserted properly logger.go:42: 08:03:00 | examples-openshift-with-htpasswd/2-check-unsecured | test step completed 2-check-unsecured logger.go:42: 08:03:00 | examples-openshift-with-htpasswd/3-check-unauthorized | starting test step 3-check-unauthorized logger.go:42: 08:03:00 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [./ensure-ingress-host.sh] logger.go:42: 08:03:00 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking the Ingress host value was populated logger.go:42: 08:03:00 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 0 logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/3-check-unauthorized | Hostname is with-htpasswd-kuttl-test-ultimate-lizard.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [sh -c JAEGER_USERNAME=wronguser JAEGER_PASSWORD=wrongpassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking an expected HTTP response logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/3-check-unauthorized | Running in OpenShift logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/3-check-unauthorized | Using Jaeger basic authentication logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 1/30 the https://with-htpasswd-kuttl-test-ultimate-lizard.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/3-check-unauthorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 2/30 the https://with-htpasswd-kuttl-test-ultimate-lizard.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/3-check-unauthorized | curl response asserted properly logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/3-check-unauthorized | test step completed 3-check-unauthorized logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/4-check-authorized | starting test step 4-check-authorized logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/4-check-authorized | running command: [./ensure-ingress-host.sh] logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/4-check-authorized | Checking the Ingress host value was populated logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/4-check-authorized | Try number 0 logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/4-check-authorized | Hostname is with-htpasswd-kuttl-test-ultimate-lizard.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/4-check-authorized | running command: [sh -c JAEGER_USERNAME=awesomeuser JAEGER_PASSWORD=awesomepassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE with-htpasswd] logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/4-check-authorized | Checking an expected HTTP response logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/4-check-authorized | Running in OpenShift logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/4-check-authorized | Using Jaeger basic authentication logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/4-check-authorized | Try number 1/30 the https://with-htpasswd-kuttl-test-ultimate-lizard.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/4-check-authorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/4-check-authorized | Try number 2/30 the https://with-htpasswd-kuttl-test-ultimate-lizard.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/4-check-authorized | curl response asserted properly logger.go:42: 08:03:01 | examples-openshift-with-htpasswd/4-check-authorized | test step completed 4-check-authorized logger.go:42: 08:03:01 | examples-openshift-with-htpasswd | examples-openshift-with-htpasswd events from ns kuttl-test-ultimate-lizard: logger.go:42: 08:03:01 | examples-openshift-with-htpasswd | 2023-12-25 08:02:47 +0000 UTC Normal Pod with-htpasswd-85bb89fcd4-gnb9v Binding Scheduled Successfully assigned kuttl-test-ultimate-lizard/with-htpasswd-85bb89fcd4-gnb9v to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:03:01 | examples-openshift-with-htpasswd | 2023-12-25 08:02:47 +0000 UTC Normal ReplicaSet.apps with-htpasswd-85bb89fcd4 SuccessfulCreate Created pod: with-htpasswd-85bb89fcd4-gnb9v replicaset-controller logger.go:42: 08:03:01 | examples-openshift-with-htpasswd | 2023-12-25 08:02:47 +0000 UTC Normal Deployment.apps with-htpasswd ScalingReplicaSet Scaled up replica set with-htpasswd-85bb89fcd4 to 1 deployment-controller logger.go:42: 08:03:01 | examples-openshift-with-htpasswd | 2023-12-25 08:02:48 +0000 UTC Warning Pod with-htpasswd-85bb89fcd4-gnb9v FailedMount MountVolume.SetUp failed for volume "with-htpasswd-ui-oauth-proxy-tls" : secret "with-htpasswd-ui-oauth-proxy-tls" not found kubelet logger.go:42: 08:03:01 | examples-openshift-with-htpasswd | 2023-12-25 08:02:48 +0000 UTC Normal Pod with-htpasswd-85bb89fcd4-gnb9v AddedInterface Add eth0 [10.129.2.52/23] from ovn-kubernetes logger.go:42: 08:03:01 | examples-openshift-with-htpasswd | 2023-12-25 08:02:48 +0000 UTC Normal Pod with-htpasswd-85bb89fcd4-gnb9v.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:03:01 | examples-openshift-with-htpasswd | 2023-12-25 08:02:48 +0000 UTC Normal Pod with-htpasswd-85bb89fcd4-gnb9v.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:03:01 | examples-openshift-with-htpasswd | 2023-12-25 08:02:48 +0000 UTC Normal Pod with-htpasswd-85bb89fcd4-gnb9v.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:03:01 | examples-openshift-with-htpasswd | 2023-12-25 08:02:48 +0000 UTC Normal Pod with-htpasswd-85bb89fcd4-gnb9v.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:03:01 | examples-openshift-with-htpasswd | 2023-12-25 08:02:48 +0000 UTC Normal Pod with-htpasswd-85bb89fcd4-gnb9v.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:03:01 | examples-openshift-with-htpasswd | 2023-12-25 08:02:48 +0000 UTC Normal Pod with-htpasswd-85bb89fcd4-gnb9v.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:03:01 | examples-openshift-with-htpasswd | Deleting namespace: kuttl-test-ultimate-lizard === CONT kuttl/harness/examples-collector-with-priority-class logger.go:42: 08:03:08 | examples-collector-with-priority-class | Creating namespace: kuttl-test-uncommon-koi logger.go:42: 08:03:08 | examples-collector-with-priority-class/0-install | starting test step 0-install logger.go:42: 08:03:08 | examples-collector-with-priority-class/0-install | PriorityClass:/collector-high-priority created logger.go:42: 08:03:08 | examples-collector-with-priority-class/0-install | Jaeger:kuttl-test-uncommon-koi/collector-with-high-priority created logger.go:42: 08:03:14 | examples-collector-with-priority-class/0-install | test step completed 0-install logger.go:42: 08:03:14 | examples-collector-with-priority-class/1-smoke-test | starting test step 1-smoke-test logger.go:42: 08:03:14 | examples-collector-with-priority-class/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE collector-with-high-priority /dev/null] logger.go:42: 08:03:15 | examples-collector-with-priority-class/1-smoke-test | Warning: resource jaegers/collector-with-high-priority is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:03:22 | examples-collector-with-priority-class/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 08:03:22 | examples-collector-with-priority-class/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:03:23 | examples-collector-with-priority-class/1-smoke-test | job.batch/report-span created logger.go:42: 08:03:23 | examples-collector-with-priority-class/1-smoke-test | job.batch/check-span created logger.go:42: 08:03:34 | examples-collector-with-priority-class/1-smoke-test | test step completed 1-smoke-test logger.go:42: 08:03:34 | examples-collector-with-priority-class | examples-collector-with-priority-class events from ns kuttl-test-uncommon-koi: logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:11 +0000 UTC Normal Pod collector-with-high-priority-66f478844f-4qlcg Binding Scheduled Successfully assigned kuttl-test-uncommon-koi/collector-with-high-priority-66f478844f-4qlcg to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:11 +0000 UTC Normal ReplicaSet.apps collector-with-high-priority-66f478844f SuccessfulCreate Created pod: collector-with-high-priority-66f478844f-4qlcg replicaset-controller logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:11 +0000 UTC Normal Deployment.apps collector-with-high-priority ScalingReplicaSet Scaled up replica set collector-with-high-priority-66f478844f to 1 deployment-controller logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:12 +0000 UTC Normal Pod collector-with-high-priority-66f478844f-4qlcg AddedInterface Add eth0 [10.129.2.53/23] from ovn-kubernetes logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:12 +0000 UTC Normal Pod collector-with-high-priority-66f478844f-4qlcg.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:12 +0000 UTC Normal Pod collector-with-high-priority-66f478844f-4qlcg.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:12 +0000 UTC Normal Pod collector-with-high-priority-66f478844f-4qlcg.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:12 +0000 UTC Normal Pod collector-with-high-priority-66f478844f-4qlcg.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:12 +0000 UTC Normal Pod collector-with-high-priority-66f478844f-4qlcg.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:12 +0000 UTC Normal Pod collector-with-high-priority-66f478844f-4qlcg.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:17 +0000 UTC Normal Pod collector-with-high-priority-66f478844f-4qlcg.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:17 +0000 UTC Normal Pod collector-with-high-priority-66f478844f-4qlcg.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:17 +0000 UTC Normal ReplicaSet.apps collector-with-high-priority-66f478844f SuccessfulDelete Deleted pod: collector-with-high-priority-66f478844f-4qlcg replicaset-controller logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:17 +0000 UTC Normal Deployment.apps collector-with-high-priority ScalingReplicaSet Scaled down replica set collector-with-high-priority-66f478844f to 0 from 1 deployment-controller logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:18 +0000 UTC Normal Pod collector-with-high-priority-746b998b86-d8l7z Binding Scheduled Successfully assigned kuttl-test-uncommon-koi/collector-with-high-priority-746b998b86-d8l7z to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:18 +0000 UTC Normal Pod collector-with-high-priority-746b998b86-d8l7z AddedInterface Add eth0 [10.129.2.54/23] from ovn-kubernetes logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:18 +0000 UTC Normal Pod collector-with-high-priority-746b998b86-d8l7z.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:18 +0000 UTC Normal Pod collector-with-high-priority-746b998b86-d8l7z.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:18 +0000 UTC Normal Pod collector-with-high-priority-746b998b86-d8l7z.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:18 +0000 UTC Normal Pod collector-with-high-priority-746b998b86-d8l7z.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:18 +0000 UTC Normal Pod collector-with-high-priority-746b998b86-d8l7z.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:18 +0000 UTC Normal Pod collector-with-high-priority-746b998b86-d8l7z.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:18 +0000 UTC Normal ReplicaSet.apps collector-with-high-priority-746b998b86 SuccessfulCreate Created pod: collector-with-high-priority-746b998b86-d8l7z replicaset-controller logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:18 +0000 UTC Normal Deployment.apps collector-with-high-priority ScalingReplicaSet Scaled up replica set collector-with-high-priority-746b998b86 to 1 deployment-controller logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:23 +0000 UTC Normal Pod check-span-lppqx Binding Scheduled Successfully assigned kuttl-test-uncommon-koi/check-span-lppqx to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:23 +0000 UTC Normal Pod check-span-lppqx AddedInterface Add eth0 [10.131.0.46/23] from ovn-kubernetes logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:23 +0000 UTC Normal Pod check-span-lppqx.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:23 +0000 UTC Normal Pod check-span-lppqx.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:23 +0000 UTC Normal Pod check-span-lppqx.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:23 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-lppqx job-controller logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:23 +0000 UTC Normal Pod report-span-9kv9z Binding Scheduled Successfully assigned kuttl-test-uncommon-koi/report-span-9kv9z to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:23 +0000 UTC Normal Pod report-span-9kv9z AddedInterface Add eth0 [10.128.2.62/23] from ovn-kubernetes logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:23 +0000 UTC Normal Pod report-span-9kv9z.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:23 +0000 UTC Normal Pod report-span-9kv9z.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:23 +0000 UTC Normal Pod report-span-9kv9z.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:23 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-9kv9z job-controller logger.go:42: 08:03:34 | examples-collector-with-priority-class | 2023-12-25 08:03:34 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:03:34 | examples-collector-with-priority-class | Deleting namespace: kuttl-test-uncommon-koi === CONT kuttl/harness/examples-agent-as-daemonset logger.go:42: 08:03:42 | examples-agent-as-daemonset | Creating namespace: kuttl-test-summary-bull logger.go:42: 08:03:42 | examples-agent-as-daemonset/0-install | starting test step 0-install logger.go:42: 08:03:42 | examples-agent-as-daemonset/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 08:03:42 | examples-agent-as-daemonset/0-install | ServiceAccount:kuttl-test-summary-bull/jaeger-agent-daemonset created logger.go:42: 08:03:42 | examples-agent-as-daemonset/0-install | test step completed 0-install logger.go:42: 08:03:42 | examples-agent-as-daemonset/1-add-policy | starting test step 1-add-policy logger.go:42: 08:03:42 | examples-agent-as-daemonset/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 08:03:42 | examples-agent-as-daemonset/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 08:03:42 | examples-agent-as-daemonset/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 08:03:47 | examples-agent-as-daemonset/1-add-policy | test step completed 1-add-policy logger.go:42: 08:03:47 | examples-agent-as-daemonset/2-install | starting test step 2-install logger.go:42: 08:03:47 | examples-agent-as-daemonset/2-install | Jaeger:kuttl-test-summary-bull/agent-as-daemonset created logger.go:42: 08:03:47 | examples-agent-as-daemonset/2-install | test step completed 2-install logger.go:42: 08:03:47 | examples-agent-as-daemonset/3- | starting test step 3- logger.go:42: 08:03:56 | examples-agent-as-daemonset/3- | test step completed 3- logger.go:42: 08:03:56 | examples-agent-as-daemonset | examples-agent-as-daemonset events from ns kuttl-test-summary-bull: logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:51 +0000 UTC Normal Pod agent-as-daemonset-bfbcc8486-gvmxw Binding Scheduled Successfully assigned kuttl-test-summary-bull/agent-as-daemonset-bfbcc8486-gvmxw to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:51 +0000 UTC Normal Pod agent-as-daemonset-bfbcc8486-gvmxw AddedInterface Add eth0 [10.129.2.55/23] from ovn-kubernetes logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:51 +0000 UTC Normal Pod agent-as-daemonset-bfbcc8486-gvmxw.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:51 +0000 UTC Normal Pod agent-as-daemonset-bfbcc8486-gvmxw.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:51 +0000 UTC Normal Pod agent-as-daemonset-bfbcc8486-gvmxw.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:51 +0000 UTC Normal Pod agent-as-daemonset-bfbcc8486-gvmxw.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:51 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-bfbcc8486 SuccessfulCreate Created pod: agent-as-daemonset-bfbcc8486-gvmxw replicaset-controller logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:51 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-bfbcc8486 to 1 deployment-controller logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:52 +0000 UTC Normal Pod agent-as-daemonset-bfbcc8486-gvmxw.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:52 +0000 UTC Normal Pod agent-as-daemonset-bfbcc8486-gvmxw.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:53 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-dx4tc Binding Scheduled Successfully assigned kuttl-test-summary-bull/agent-as-daemonset-agent-daemonset-dx4tc to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:53 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-dx4tc AddedInterface Add eth0 [10.131.0.47/23] from ovn-kubernetes logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:53 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-dx4tc.spec.containers{jaeger-agent-daemonset} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:53 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-wbhjj Binding Scheduled Successfully assigned kuttl-test-summary-bull/agent-as-daemonset-agent-daemonset-wbhjj to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:53 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-wbhjj AddedInterface Add eth0 [10.128.2.63/23] from ovn-kubernetes logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:53 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-zktq5 Binding Scheduled Successfully assigned kuttl-test-summary-bull/agent-as-daemonset-agent-daemonset-zktq5 to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:53 +0000 UTC Normal DaemonSet.apps agent-as-daemonset-agent-daemonset SuccessfulCreate Created pod: agent-as-daemonset-agent-daemonset-dx4tc daemonset-controller logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:53 +0000 UTC Normal DaemonSet.apps agent-as-daemonset-agent-daemonset SuccessfulCreate Created pod: agent-as-daemonset-agent-daemonset-wbhjj daemonset-controller logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:53 +0000 UTC Normal DaemonSet.apps agent-as-daemonset-agent-daemonset SuccessfulCreate Created pod: agent-as-daemonset-agent-daemonset-zktq5 daemonset-controller logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:54 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-dx4tc.spec.containers{jaeger-agent-daemonset} Created Created container jaeger-agent-daemonset kubelet logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:54 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-dx4tc.spec.containers{jaeger-agent-daemonset} Started Started container jaeger-agent-daemonset kubelet logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:54 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-wbhjj.spec.containers{jaeger-agent-daemonset} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" kubelet logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:54 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-zktq5 AddedInterface Add eth0 [10.129.2.56/23] from ovn-kubernetes logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:54 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-zktq5.spec.containers{jaeger-agent-daemonset} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:54 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-zktq5.spec.containers{jaeger-agent-daemonset} Created Created container jaeger-agent-daemonset kubelet logger.go:42: 08:03:56 | examples-agent-as-daemonset | 2023-12-25 08:03:54 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-zktq5.spec.containers{jaeger-agent-daemonset} Started Started container jaeger-agent-daemonset kubelet logger.go:42: 08:03:56 | examples-agent-as-daemonset | Deleting namespace: kuttl-test-summary-bull === CONT kuttl/harness/examples-all-in-one-with-options logger.go:42: 08:04:02 | examples-all-in-one-with-options | Creating namespace: kuttl-test-unique-dog logger.go:42: 08:04:02 | examples-all-in-one-with-options/0-install | starting test step 0-install logger.go:42: 08:04:02 | examples-all-in-one-with-options/0-install | Jaeger:kuttl-test-unique-dog/my-jaeger created logger.go:42: 08:04:09 | examples-all-in-one-with-options/0-install | test step completed 0-install logger.go:42: 08:04:09 | examples-all-in-one-with-options/1-smoke-test | starting test step 1-smoke-test logger.go:42: 08:04:09 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 08:04:10 | examples-all-in-one-with-options/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:04:16 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443/jaeger MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 08:04:17 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:04:17 | examples-all-in-one-with-options/1-smoke-test | job.batch/report-span created logger.go:42: 08:04:17 | examples-all-in-one-with-options/1-smoke-test | job.batch/check-span created logger.go:42: 08:04:28 | examples-all-in-one-with-options/1-smoke-test | test step completed 1-smoke-test logger.go:42: 08:04:28 | examples-all-in-one-with-options | examples-all-in-one-with-options events from ns kuttl-test-unique-dog: logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:06 +0000 UTC Normal Pod my-jaeger-75cc6b56b5-fkf6t Binding Scheduled Successfully assigned kuttl-test-unique-dog/my-jaeger-75cc6b56b5-fkf6t to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:06 +0000 UTC Normal Pod my-jaeger-75cc6b56b5-fkf6t AddedInterface Add eth0 [10.129.2.57/23] from ovn-kubernetes logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:06 +0000 UTC Normal Pod my-jaeger-75cc6b56b5-fkf6t.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:06 +0000 UTC Normal Pod my-jaeger-75cc6b56b5-fkf6t.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:06 +0000 UTC Normal Pod my-jaeger-75cc6b56b5-fkf6t.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:06 +0000 UTC Normal Pod my-jaeger-75cc6b56b5-fkf6t.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:06 +0000 UTC Normal Pod my-jaeger-75cc6b56b5-fkf6t.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:06 +0000 UTC Normal Pod my-jaeger-75cc6b56b5-fkf6t.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:06 +0000 UTC Normal ReplicaSet.apps my-jaeger-75cc6b56b5 SuccessfulCreate Created pod: my-jaeger-75cc6b56b5-fkf6t replicaset-controller logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:06 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-75cc6b56b5 to 1 deployment-controller logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:11 +0000 UTC Normal Pod my-jaeger-75cc6b56b5-fkf6t.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:11 +0000 UTC Normal Pod my-jaeger-75cc6b56b5-fkf6t.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:11 +0000 UTC Normal ReplicaSet.apps my-jaeger-75cc6b56b5 SuccessfulDelete Deleted pod: my-jaeger-75cc6b56b5-fkf6t replicaset-controller logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:11 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-75cc6b56b5 to 0 from 1 deployment-controller logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:12 +0000 UTC Normal Pod my-jaeger-7bb7b59846-8ftgp Binding Scheduled Successfully assigned kuttl-test-unique-dog/my-jaeger-7bb7b59846-8ftgp to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:12 +0000 UTC Normal ReplicaSet.apps my-jaeger-7bb7b59846 SuccessfulCreate Created pod: my-jaeger-7bb7b59846-8ftgp replicaset-controller logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:12 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-7bb7b59846 to 1 deployment-controller logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:13 +0000 UTC Normal Pod my-jaeger-7bb7b59846-8ftgp AddedInterface Add eth0 [10.129.2.58/23] from ovn-kubernetes logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:13 +0000 UTC Normal Pod my-jaeger-7bb7b59846-8ftgp.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:13 +0000 UTC Normal Pod my-jaeger-7bb7b59846-8ftgp.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:13 +0000 UTC Normal Pod my-jaeger-7bb7b59846-8ftgp.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:13 +0000 UTC Normal Pod my-jaeger-7bb7b59846-8ftgp.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:13 +0000 UTC Normal Pod my-jaeger-7bb7b59846-8ftgp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:13 +0000 UTC Normal Pod my-jaeger-7bb7b59846-8ftgp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:17 +0000 UTC Normal Pod check-span-b86jz Binding Scheduled Successfully assigned kuttl-test-unique-dog/check-span-b86jz to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:17 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-b86jz job-controller logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:17 +0000 UTC Normal Pod report-span-mmdfz Binding Scheduled Successfully assigned kuttl-test-unique-dog/report-span-mmdfz to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:17 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-mmdfz job-controller logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:18 +0000 UTC Normal Pod check-span-b86jz AddedInterface Add eth0 [10.131.0.48/23] from ovn-kubernetes logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:18 +0000 UTC Normal Pod check-span-b86jz.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:18 +0000 UTC Normal Pod check-span-b86jz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:18 +0000 UTC Normal Pod check-span-b86jz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:18 +0000 UTC Normal Pod report-span-mmdfz AddedInterface Add eth0 [10.128.2.64/23] from ovn-kubernetes logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:18 +0000 UTC Normal Pod report-span-mmdfz.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:18 +0000 UTC Normal Pod report-span-mmdfz.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:18 +0000 UTC Normal Pod report-span-mmdfz.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:04:28 | examples-all-in-one-with-options | 2023-12-25 08:04:28 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:04:28 | examples-all-in-one-with-options | Deleting namespace: kuttl-test-unique-dog === CONT kuttl/harness/examples-agent-with-priority-class logger.go:42: 08:04:36 | examples-agent-with-priority-class | Creating namespace: kuttl-test-knowing-sunbeam logger.go:42: 08:04:36 | examples-agent-with-priority-class/0-install | starting test step 0-install logger.go:42: 08:04:36 | examples-agent-with-priority-class/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 08:04:36 | examples-agent-with-priority-class/0-install | ServiceAccount:kuttl-test-knowing-sunbeam/jaeger-agent-daemonset created logger.go:42: 08:04:36 | examples-agent-with-priority-class/0-install | test step completed 0-install logger.go:42: 08:04:36 | examples-agent-with-priority-class/1-add-policy | starting test step 1-add-policy logger.go:42: 08:04:36 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 08:04:36 | examples-agent-with-priority-class/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 08:04:36 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 08:04:42 | examples-agent-with-priority-class/1-add-policy | test step completed 1-add-policy logger.go:42: 08:04:42 | examples-agent-with-priority-class/2-install | starting test step 2-install logger.go:42: 08:04:42 | examples-agent-with-priority-class/2-install | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 08:04:49 | examples-agent-with-priority-class/2-install | Error from server (NotFound): deployments.apps "agent-as-daemonset" not found logger.go:42: 08:04:49 | examples-agent-with-priority-class/2-install | command failure, skipping 2 additional commands logger.go:42: 08:04:49 | examples-agent-with-priority-class/2-install | PriorityClass:/high-priority created logger.go:42: 08:04:49 | examples-agent-with-priority-class/2-install | Jaeger:kuttl-test-knowing-sunbeam/agent-as-daemonset updated case.go:364: failed in step 2-install case.go:366: exit status 1 logger.go:42: 08:04:49 | examples-agent-with-priority-class | examples-agent-with-priority-class events from ns kuttl-test-knowing-sunbeam: logger.go:42: 08:04:49 | examples-agent-with-priority-class | Deleting namespace: kuttl-test-knowing-sunbeam === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (1187.33s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.41s) --- PASS: kuttl/harness/examples-simple-prod (69.24s) --- PASS: kuttl/harness/examples-with-sampling (57.12s) --- PASS: kuttl/harness/examples-with-cassandra (55.05s) --- FAIL: kuttl/harness/examples-with-badger-and-volume (622.47s) --- PASS: kuttl/harness/examples-with-badger (38.97s) --- PASS: kuttl/harness/examples-simplest (39.06s) --- PASS: kuttl/harness/examples-simple-prod-with-volumes (71.04s) --- PASS: kuttl/harness/examples-business-application-injected-sidecar (38.37s) --- PASS: kuttl/harness/examples-service-types (57.18s) --- PASS: kuttl/harness/examples-openshift-with-htpasswd (23.97s) --- PASS: kuttl/harness/examples-collector-with-priority-class (34.42s) --- PASS: kuttl/harness/examples-agent-as-daemonset (20.21s) --- PASS: kuttl/harness/examples-all-in-one-with-options (33.95s) --- FAIL: kuttl/harness/examples-agent-with-priority-class (19.70s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name examples --report --output /logs/artifacts/examples.xml ./artifacts/kuttl-report.xml time="2023-12-25T08:04:56Z" level=debug msg="Setting a new name for the test suites" time="2023-12-25T08:04:56Z" level=debug msg="Removing 'artifacts' TestCase" time="2023-12-25T08:04:56Z" level=debug msg="normalizing test case names" time="2023-12-25T08:04:56Z" level=debug msg="examples/artifacts -> examples_artifacts" time="2023-12-25T08:04:56Z" level=debug msg="examples/examples-simple-prod -> examples_examples_simple_prod" time="2023-12-25T08:04:56Z" level=debug msg="examples/examples-with-sampling -> examples_examples_with_sampling" time="2023-12-25T08:04:56Z" level=debug msg="examples/examples-with-cassandra -> examples_examples_with_cassandra" time="2023-12-25T08:04:56Z" level=debug msg="examples/examples-with-badger-and-volume -> examples_examples_with_badger_and_volume" time="2023-12-25T08:04:56Z" level=debug msg="examples/examples-with-badger -> examples_examples_with_badger" time="2023-12-25T08:04:56Z" level=debug msg="examples/examples-simplest -> examples_examples_simplest" time="2023-12-25T08:04:56Z" level=debug msg="examples/examples-simple-prod-with-volumes -> examples_examples_simple_prod_with_volumes" time="2023-12-25T08:04:56Z" level=debug msg="examples/examples-business-application-injected-sidecar -> examples_examples_business_application_injected_sidecar" time="2023-12-25T08:04:56Z" level=debug msg="examples/examples-service-types -> examples_examples_service_types" time="2023-12-25T08:04:56Z" level=debug msg="examples/examples-openshift-with-htpasswd -> examples_examples_openshift_with_htpasswd" time="2023-12-25T08:04:56Z" level=debug msg="examples/examples-collector-with-priority-class -> examples_examples_collector_with_priority_class" time="2023-12-25T08:04:56Z" level=debug msg="examples/examples-agent-as-daemonset -> examples_examples_agent_as_daemonset" time="2023-12-25T08:04:56Z" level=debug msg="examples/examples-all-in-one-with-options -> examples_examples_all_in_one_with_options" time="2023-12-25T08:04:56Z" level=debug msg="examples/examples-agent-with-priority-class -> examples_examples_agent_with_priority_class" +---------------------------------------------------------+--------+ | NAME | RESULT | +---------------------------------------------------------+--------+ | examples_artifacts | passed | | examples_examples_simple_prod | passed | | examples_examples_with_sampling | passed | | examples_examples_with_cassandra | passed | | examples_examples_with_badger_and_volume | failed | | examples_examples_with_badger | passed | | examples_examples_simplest | passed | | examples_examples_simple_prod_with_volumes | passed | | examples_examples_business_application_injected_sidecar | passed | | examples_examples_service_types | passed | | examples_examples_openshift_with_htpasswd | passed | | examples_examples_collector_with_priority_class | passed | | examples_examples_agent_as_daemonset | passed | | examples_examples_all_in_one_with_options | passed | | examples_examples_agent_with_priority_class | failed | +---------------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh generate false true + '[' 3 -ne 3 ']' + test_suite_name=generate + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/generate.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-generate make[2]: Entering directory '/tmp/jaeger-tests' test -s /tmp/jaeger-tests/bin/operator-sdk || curl -sLo /tmp/jaeger-tests/bin/operator-sdk https://github.com/operator-framework/operator-sdk/releases/download/v1.32.0/operator-sdk_`go env GOOS`_`go env GOARCH` ./hack/install/install-golangci-lint.sh Installing golangci-lint golangci-lint 1.55.2 is installed already ./hack/install/install-goimports.sh Installing goimports Try 0... go install golang.org/x/tools/cmd/goimports@v0.1.12 >>>> Formatting code... ./.ci/format.sh >>>> Building... ./hack/install/install-dependencies.sh Installing go dependencies Try 0... go mod download GOOS= GOARCH= CGO_ENABLED=0 GO111MODULE=on go build -ldflags "-X "github.com/jaegertracing/jaeger-operator/pkg/version".version="1.52.0" -X "github.com/jaegertracing/jaeger-operator/pkg/version".buildDate=2023-12-25T08:04:57Z -X "github.com/jaegertracing/jaeger-operator/pkg/version".defaultJaeger="1.52.0"" -o "bin/jaeger-operator" main.go JAEGER_VERSION="1.52.0" ./tests/e2e/generate/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-12-22-165425 True False 39m Cluster version is 4.15.0-0.nightly-2023-12-22-165425' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-12-22-165425 True False 39m Cluster version is 4.15.0-0.nightly-2023-12-22-165425' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/generate/render.sh ++ export SUITE_DIR=./tests/e2e/generate ++ SUITE_DIR=./tests/e2e/generate ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/generate ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test generate 'This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 2 -ne 2 ']' + test_name=generate + message='This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/generate/_build + '[' _build '!=' _build ']' + rm -rf generate + warning 'generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed\e[0m' WAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running generate E2E tests' Running generate E2E tests + cd tests/e2e/generate/_build + set +e + KUBECONFIG=/tmp/kubeconfig-1441553849 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 08:05:12 | artifacts | Creating namespace: kuttl-test-grand-spider logger.go:42: 08:05:12 | artifacts | artifacts events from ns kuttl-test-grand-spider: logger.go:42: 08:05:12 | artifacts | Deleting namespace: kuttl-test-grand-spider === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.36s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.20s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name generate --report --output /logs/artifacts/generate.xml ./artifacts/kuttl-report.xml time="2023-12-25T08:05:19Z" level=debug msg="Setting a new name for the test suites" time="2023-12-25T08:05:19Z" level=debug msg="Removing 'artifacts' TestCase" time="2023-12-25T08:05:19Z" level=debug msg="normalizing test case names" time="2023-12-25T08:05:19Z" level=debug msg="generate/artifacts -> generate_artifacts" +--------------------+--------+ | NAME | RESULT | +--------------------+--------+ | generate_artifacts | passed | +--------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh miscellaneous false true + '[' 3 -ne 3 ']' + test_suite_name=miscellaneous + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/miscellaneous.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-miscellaneous make[2]: Entering directory '/tmp/jaeger-tests' SKIP_ES_EXTERNAL=true ./tests/e2e/miscellaneous/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-12-22-165425 True False 39m Cluster version is 4.15.0-0.nightly-2023-12-22-165425' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-12-22-165425 True False 39m Cluster version is 4.15.0-0.nightly-2023-12-22-165425' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/miscellaneous/render.sh ++ export SUITE_DIR=./tests/e2e/miscellaneous ++ SUITE_DIR=./tests/e2e/miscellaneous ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/miscellaneous ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test cassandra-spark 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=cassandra-spark + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf cassandra-spark + warning 'cassandra-spark: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: cassandra-spark: Test not supported in OpenShift\e[0m' WAR: cassandra-spark: Test not supported in OpenShift + start_test collector-autoscale + '[' 1 -ne 1 ']' + test_name=collector-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-autoscale\e[0m' Rendering files for test collector-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p collector-autoscale + cd collector-autoscale + jaeger_name=simple-prod + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + ELASTICSEARCH_NODECOUNT=1 + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.resources.requests.memory="200m"' 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.autoscale=true 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.minReplicas=1 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.maxReplicas=2 01-install.yaml + version_lt 1.27 1.23 ++ echo 1.27 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.27 '!=' 1.27 + rm ./03-assert.yaml + generate_otlp_e2e_tests http + test_protocol=http + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-http\e[0m' Rendering files for test collector-otlp-allinone-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-autoscale + '[' collector-autoscale '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-http + cd collector-otlp-allinone-http + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger http true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-http\e[0m' Rendering files for test collector-otlp-production-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-http + '[' collector-otlp-allinone-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-http + cd collector-otlp-production-http + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger http true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + generate_otlp_e2e_tests grpc + test_protocol=grpc + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-grpc\e[0m' Rendering files for test collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-http + '[' collector-otlp-production-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-grpc + cd collector-otlp-allinone-grpc + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-grpc\e[0m' Rendering files for test collector-otlp-production-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-grpc + '[' collector-otlp-allinone-grpc '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-grpc + cd collector-otlp-production-grpc + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + '[' true = true ']' + skip_test istio 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=istio + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-grpc + '[' collector-otlp-production-grpc '!=' _build ']' + cd .. + rm -rf istio + warning 'istio: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: istio: Test not supported in OpenShift\e[0m' WAR: istio: Test not supported in OpenShift + '[' true = true ']' + skip_test outside-cluster 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=outside-cluster + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf outside-cluster + warning 'outside-cluster: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: outside-cluster: Test not supported in OpenShift\e[0m' WAR: outside-cluster: Test not supported in OpenShift + start_test set-custom-img + '[' 1 -ne 1 ']' + test_name=set-custom-img + echo =========================================================================== =========================================================================== + info 'Rendering files for test set-custom-img' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test set-custom-img\e[0m' Rendering files for test set-custom-img + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p set-custom-img + cd set-custom-img + jaeger_name=my-jaeger + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.image="test"' ./02-install.yaml + '[' true = true ']' + skip_test non-cluster-wide 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=non-cluster-wide + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/set-custom-img + '[' set-custom-img '!=' _build ']' + cd .. + rm -rf non-cluster-wide + warning 'non-cluster-wide: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: non-cluster-wide: Test not supported in OpenShift\e[0m' WAR: non-cluster-wide: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running miscellaneous E2E tests' Running miscellaneous E2E tests + cd tests/e2e/miscellaneous/_build + set +e + KUBECONFIG=/tmp/kubeconfig-1441553849 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/collector-autoscale === PAUSE kuttl/harness/collector-autoscale === RUN kuttl/harness/collector-otlp-allinone-grpc === PAUSE kuttl/harness/collector-otlp-allinone-grpc === RUN kuttl/harness/collector-otlp-allinone-http === PAUSE kuttl/harness/collector-otlp-allinone-http === RUN kuttl/harness/collector-otlp-production-grpc === PAUSE kuttl/harness/collector-otlp-production-grpc === RUN kuttl/harness/collector-otlp-production-http === PAUSE kuttl/harness/collector-otlp-production-http === RUN kuttl/harness/set-custom-img === PAUSE kuttl/harness/set-custom-img === CONT kuttl/harness/artifacts logger.go:42: 08:05:29 | artifacts | Creating namespace: kuttl-test-tight-escargot logger.go:42: 08:05:29 | artifacts | artifacts events from ns kuttl-test-tight-escargot: logger.go:42: 08:05:29 | artifacts | Deleting namespace: kuttl-test-tight-escargot === CONT kuttl/harness/collector-otlp-production-grpc logger.go:42: 08:05:36 | collector-otlp-production-grpc | Creating namespace: kuttl-test-thankful-goat logger.go:42: 08:05:36 | collector-otlp-production-grpc/1-install | starting test step 1-install logger.go:42: 08:05:36 | collector-otlp-production-grpc/1-install | Jaeger:kuttl-test-thankful-goat/my-jaeger created logger.go:42: 08:06:11 | collector-otlp-production-grpc/1-install | test step completed 1-install logger.go:42: 08:06:11 | collector-otlp-production-grpc/2-smoke-test | starting test step 2-smoke-test logger.go:42: 08:06:11 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 08:06:12 | collector-otlp-production-grpc/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:06:19 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 08:06:19 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:06:19 | collector-otlp-production-grpc/2-smoke-test | job.batch/report-span created logger.go:42: 08:06:19 | collector-otlp-production-grpc/2-smoke-test | job.batch/check-span created logger.go:42: 08:06:39 | collector-otlp-production-grpc/2-smoke-test | test step completed 2-smoke-test logger.go:42: 08:06:39 | collector-otlp-production-grpc | collector-otlp-production-grpc events from ns kuttl-test-thankful-goat: logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:05:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestthankfulgoatmyjaeger-1-77fcb675k9bhh Binding Scheduled Successfully assigned kuttl-test-thankful-goat/elasticsearch-cdm-kuttltestthankfulgoatmyjaeger-1-77fcb675k9bhh to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:05:41 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestthankfulgoatmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestthankfulgoatmyjaeger-1-77fcb675d9 to 1 deployment-controller logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:05:42 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestthankfulgoatmyjaeger-1-77fcb675d9 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestthankfulgoatmyjaeger-1-77fcb675k9bhh replicaset-controller logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:05:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestthankfulgoatmyjaeger-1-77fcb675k9bhh AddedInterface Add eth0 [10.129.2.59/23] from ovn-kubernetes logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:05:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestthankfulgoatmyjaeger-1-77fcb675k9bhh.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:05:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestthankfulgoatmyjaeger-1-77fcb675k9bhh.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:05:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestthankfulgoatmyjaeger-1-77fcb675k9bhh.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:05:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestthankfulgoatmyjaeger-1-77fcb675k9bhh.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:05:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestthankfulgoatmyjaeger-1-77fcb675k9bhh.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:05:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestthankfulgoatmyjaeger-1-77fcb675k9bhh.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:05:52 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestthankfulgoatmyjaeger-1-77fcb675k9bhh.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:05:57 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestthankfulgoatmyjaeger-1-77fcb675k9bhh.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:08 +0000 UTC Normal Pod my-jaeger-collector-5988569d78-fm42w Binding Scheduled Successfully assigned kuttl-test-thankful-goat/my-jaeger-collector-5988569d78-fm42w to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:08 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5988569d78 SuccessfulCreate Created pod: my-jaeger-collector-5988569d78-fm42w replicaset-controller logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:08 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-5988569d78 to 1 deployment-controller logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:08 +0000 UTC Normal Pod my-jaeger-query-86b4fc5574-rd5tk Binding Scheduled Successfully assigned kuttl-test-thankful-goat/my-jaeger-query-86b4fc5574-rd5tk to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:08 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-86b4fc5574 SuccessfulCreate Created pod: my-jaeger-query-86b4fc5574-rd5tk replicaset-controller logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:08 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-86b4fc5574 to 1 deployment-controller logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:09 +0000 UTC Normal Pod my-jaeger-collector-5988569d78-fm42w AddedInterface Add eth0 [10.128.2.65/23] from ovn-kubernetes logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:09 +0000 UTC Normal Pod my-jaeger-collector-5988569d78-fm42w.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:09 +0000 UTC Normal Pod my-jaeger-collector-5988569d78-fm42w.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:09 +0000 UTC Normal Pod my-jaeger-collector-5988569d78-fm42w.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:09 +0000 UTC Normal Pod my-jaeger-query-86b4fc5574-rd5tk AddedInterface Add eth0 [10.131.0.49/23] from ovn-kubernetes logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:09 +0000 UTC Normal Pod my-jaeger-query-86b4fc5574-rd5tk.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:09 +0000 UTC Normal Pod my-jaeger-query-86b4fc5574-rd5tk.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:09 +0000 UTC Normal Pod my-jaeger-query-86b4fc5574-rd5tk.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:09 +0000 UTC Normal Pod my-jaeger-query-86b4fc5574-rd5tk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:09 +0000 UTC Normal Pod my-jaeger-query-86b4fc5574-rd5tk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:09 +0000 UTC Normal Pod my-jaeger-query-86b4fc5574-rd5tk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:09 +0000 UTC Normal Pod my-jaeger-query-86b4fc5574-rd5tk.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:09 +0000 UTC Normal Pod my-jaeger-query-86b4fc5574-rd5tk.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:09 +0000 UTC Normal Pod my-jaeger-query-86b4fc5574-rd5tk.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:15 +0000 UTC Normal Pod my-jaeger-query-86b4fc5574-rd5tk.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:15 +0000 UTC Normal Pod my-jaeger-query-86b4fc5574-rd5tk.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:15 +0000 UTC Normal Pod my-jaeger-query-86b4fc5574-rd5tk.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:15 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-86b4fc5574 SuccessfulDelete Deleted pod: my-jaeger-query-86b4fc5574-rd5tk replicaset-controller logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:15 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-86b4fc5574 to 0 from 1 deployment-controller logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:16 +0000 UTC Normal Pod my-jaeger-query-ff6dc8959-glgkc Binding Scheduled Successfully assigned kuttl-test-thankful-goat/my-jaeger-query-ff6dc8959-glgkc to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:16 +0000 UTC Normal Pod my-jaeger-query-ff6dc8959-glgkc AddedInterface Add eth0 [10.131.0.50/23] from ovn-kubernetes logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:16 +0000 UTC Normal Pod my-jaeger-query-ff6dc8959-glgkc.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:16 +0000 UTC Normal Pod my-jaeger-query-ff6dc8959-glgkc.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:16 +0000 UTC Normal Pod my-jaeger-query-ff6dc8959-glgkc.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:16 +0000 UTC Normal Pod my-jaeger-query-ff6dc8959-glgkc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:16 +0000 UTC Normal Pod my-jaeger-query-ff6dc8959-glgkc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:16 +0000 UTC Normal Pod my-jaeger-query-ff6dc8959-glgkc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:16 +0000 UTC Normal Pod my-jaeger-query-ff6dc8959-glgkc.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:16 +0000 UTC Normal Pod my-jaeger-query-ff6dc8959-glgkc.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:16 +0000 UTC Normal Pod my-jaeger-query-ff6dc8959-glgkc.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:16 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-ff6dc8959 SuccessfulCreate Created pod: my-jaeger-query-ff6dc8959-glgkc replicaset-controller logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:16 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-ff6dc8959 to 1 deployment-controller logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:19 +0000 UTC Normal Pod check-span-6gpmp Binding Scheduled Successfully assigned kuttl-test-thankful-goat/check-span-6gpmp to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:19 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-6gpmp job-controller logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:19 +0000 UTC Normal Pod report-span-n6lm8 Binding Scheduled Successfully assigned kuttl-test-thankful-goat/report-span-n6lm8 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:19 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-n6lm8 job-controller logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:20 +0000 UTC Normal Pod check-span-6gpmp AddedInterface Add eth0 [10.128.2.67/23] from ovn-kubernetes logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:20 +0000 UTC Normal Pod check-span-6gpmp.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:20 +0000 UTC Normal Pod check-span-6gpmp.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:20 +0000 UTC Normal Pod check-span-6gpmp.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:20 +0000 UTC Normal Pod report-span-n6lm8 AddedInterface Add eth0 [10.128.2.66/23] from ovn-kubernetes logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:20 +0000 UTC Normal Pod report-span-n6lm8.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:20 +0000 UTC Normal Pod report-span-n6lm8.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:20 +0000 UTC Normal Pod report-span-n6lm8.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:26 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:26 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:26 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:06:39 | collector-otlp-production-grpc | 2023-12-25 08:06:38 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:06:39 | collector-otlp-production-grpc | Deleting namespace: kuttl-test-thankful-goat === CONT kuttl/harness/set-custom-img logger.go:42: 08:06:47 | set-custom-img | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:06:47 | set-custom-img | Ignoring check-collector-img.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:06:47 | set-custom-img | Creating namespace: kuttl-test-main-gnu logger.go:42: 08:06:47 | set-custom-img/1-install | starting test step 1-install logger.go:42: 08:06:47 | set-custom-img/1-install | Jaeger:kuttl-test-main-gnu/my-jaeger created logger.go:42: 08:07:25 | set-custom-img/1-install | test step completed 1-install logger.go:42: 08:07:25 | set-custom-img/2-install | starting test step 2-install logger.go:42: 08:07:25 | set-custom-img/2-install | Jaeger:kuttl-test-main-gnu/my-jaeger updated logger.go:42: 08:07:25 | set-custom-img/2-install | test step completed 2-install logger.go:42: 08:07:25 | set-custom-img/3-check-image | starting test step 3-check-image logger.go:42: 08:07:25 | set-custom-img/3-check-image | running command: [sh -c ./check-collector-img.sh] logger.go:42: 08:07:25 | set-custom-img/3-check-image | Collector image missmatch. Expected: test. Has: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9 logger.go:42: 08:07:31 | set-custom-img/3-check-image | Collector image asserted properly! logger.go:42: 08:07:31 | set-custom-img/3-check-image | test step completed 3-check-image logger.go:42: 08:07:31 | set-custom-img | set-custom-img events from ns kuttl-test-main-gnu: logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:06:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaingnumyjaeger-1-66d4697d4f-npvlz Binding Scheduled Successfully assigned kuttl-test-main-gnu/elasticsearch-cdm-kuttltestmaingnumyjaeger-1-66d4697d4f-npvlz to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:06:53 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestmaingnumyjaeger-1-66d4697d4f SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestmaingnumyjaeger-1-66d4697d4f-npvlz replicaset-controller logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:06:53 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestmaingnumyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestmaingnumyjaeger-1-66d4697d4f to 1 deployment-controller logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:06:54 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmaingnumyjaeger-1-66d4697d4f-npvlz FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:06:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaingnumyjaeger-1-66d4697d4f-npvlz AddedInterface Add eth0 [10.129.2.60/23] from ovn-kubernetes logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:06:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaingnumyjaeger-1-66d4697d4f-npvlz.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:06:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaingnumyjaeger-1-66d4697d4f-npvlz.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:06:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaingnumyjaeger-1-66d4697d4f-npvlz.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:06:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaingnumyjaeger-1-66d4697d4f-npvlz.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:06:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaingnumyjaeger-1-66d4697d4f-npvlz.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:06:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaingnumyjaeger-1-66d4697d4f-npvlz.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:10 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmaingnumyjaeger-1-66d4697d4f-npvlz.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:20 +0000 UTC Normal Pod my-jaeger-collector-7ccbb9c869-mwccb Binding Scheduled Successfully assigned kuttl-test-main-gnu/my-jaeger-collector-7ccbb9c869-mwccb to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:20 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7ccbb9c869 SuccessfulCreate Created pod: my-jaeger-collector-7ccbb9c869-mwccb replicaset-controller logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:20 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7ccbb9c869 to 1 deployment-controller logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:20 +0000 UTC Normal Pod my-jaeger-query-ff8f5cb4-xpkmg Binding Scheduled Successfully assigned kuttl-test-main-gnu/my-jaeger-query-ff8f5cb4-xpkmg to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:20 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-ff8f5cb4 SuccessfulCreate Created pod: my-jaeger-query-ff8f5cb4-xpkmg replicaset-controller logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:20 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-ff8f5cb4 to 1 deployment-controller logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:21 +0000 UTC Normal Pod my-jaeger-collector-7ccbb9c869-mwccb AddedInterface Add eth0 [10.128.2.68/23] from ovn-kubernetes logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:21 +0000 UTC Normal Pod my-jaeger-collector-7ccbb9c869-mwccb.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:21 +0000 UTC Normal Pod my-jaeger-collector-7ccbb9c869-mwccb.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:21 +0000 UTC Normal Pod my-jaeger-collector-7ccbb9c869-mwccb.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:22 +0000 UTC Warning Pod my-jaeger-query-ff8f5cb4-xpkmg FailedMount MountVolume.SetUp failed for volume "certs" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:22 +0000 UTC Normal Pod my-jaeger-query-ff8f5cb4-xpkmg AddedInterface Add eth0 [10.131.0.51/23] from ovn-kubernetes logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:22 +0000 UTC Normal Pod my-jaeger-query-ff8f5cb4-xpkmg.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:22 +0000 UTC Normal Pod my-jaeger-query-ff8f5cb4-xpkmg.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:22 +0000 UTC Normal Pod my-jaeger-query-ff8f5cb4-xpkmg.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:22 +0000 UTC Normal Pod my-jaeger-query-ff8f5cb4-xpkmg.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:22 +0000 UTC Normal Pod my-jaeger-query-ff8f5cb4-xpkmg.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:22 +0000 UTC Normal Pod my-jaeger-query-ff8f5cb4-xpkmg.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:22 +0000 UTC Normal Pod my-jaeger-query-ff8f5cb4-xpkmg.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:23 +0000 UTC Normal Pod my-jaeger-query-ff8f5cb4-xpkmg.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:23 +0000 UTC Normal Pod my-jaeger-query-ff8f5cb4-xpkmg.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:29 +0000 UTC Normal Pod my-jaeger-collector-79c49b5d6c-8wpwl Binding Scheduled Successfully assigned kuttl-test-main-gnu/my-jaeger-collector-79c49b5d6c-8wpwl to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:29 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-79c49b5d6c SuccessfulCreate Created pod: my-jaeger-collector-79c49b5d6c-8wpwl replicaset-controller logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:29 +0000 UTC Normal Pod my-jaeger-collector-7ccbb9c869-mwccb.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:29 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7ccbb9c869 SuccessfulDelete Deleted pod: my-jaeger-collector-7ccbb9c869-mwccb replicaset-controller logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:29 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-7ccbb9c869 to 0 from 1 deployment-controller logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:29 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-79c49b5d6c to 1 deployment-controller logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:30 +0000 UTC Normal Pod my-jaeger-collector-79c49b5d6c-8wpwl AddedInterface Add eth0 [10.128.2.69/23] from ovn-kubernetes logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:30 +0000 UTC Normal Pod my-jaeger-collector-79c49b5d6c-8wpwl.spec.containers{jaeger-collector} Pulling Pulling image "test" kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:30 +0000 UTC Warning Pod my-jaeger-collector-79c49b5d6c-8wpwl.spec.containers{jaeger-collector} Failed Failed to pull image "test": reading manifest latest in docker.io/library/test: requested access to the resource is denied kubelet logger.go:42: 08:07:31 | set-custom-img | 2023-12-25 08:07:30 +0000 UTC Warning Pod my-jaeger-collector-79c49b5d6c-8wpwl.spec.containers{jaeger-collector} Failed Error: ErrImagePull kubelet logger.go:42: 08:07:31 | set-custom-img | Deleting namespace: kuttl-test-main-gnu === CONT kuttl/harness/collector-otlp-production-http logger.go:42: 08:07:37 | collector-otlp-production-http | Creating namespace: kuttl-test-allowed-skunk logger.go:42: 08:07:37 | collector-otlp-production-http/1-install | starting test step 1-install logger.go:42: 08:07:37 | collector-otlp-production-http/1-install | Jaeger:kuttl-test-allowed-skunk/my-jaeger created logger.go:42: 08:08:15 | collector-otlp-production-http/1-install | test step completed 1-install logger.go:42: 08:08:15 | collector-otlp-production-http/2-smoke-test | starting test step 2-smoke-test logger.go:42: 08:08:15 | collector-otlp-production-http/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 08:08:16 | collector-otlp-production-http/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:08:22 | collector-otlp-production-http/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 08:08:23 | collector-otlp-production-http/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:08:23 | collector-otlp-production-http/2-smoke-test | job.batch/report-span created logger.go:42: 08:08:23 | collector-otlp-production-http/2-smoke-test | job.batch/check-span created logger.go:42: 08:08:35 | collector-otlp-production-http/2-smoke-test | test step completed 2-smoke-test logger.go:42: 08:08:35 | collector-otlp-production-http | collector-otlp-production-http events from ns kuttl-test-allowed-skunk: logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:07:44 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestallowedskunkmyjaeger-1-777866cbc7 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestallowedskunkmyjaeger-1-777866cbm5ndz replicaset-controller logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:07:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestallowedskunkmyjaeger-1-777866cbm5ndz Binding Scheduled Successfully assigned kuttl-test-allowed-skunk/elasticsearch-cdm-kuttltestallowedskunkmyjaeger-1-777866cbm5ndz to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:07:44 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestallowedskunkmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestallowedskunkmyjaeger-1-777866cbc7 to 1 deployment-controller logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:07:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestallowedskunkmyjaeger-1-777866cbm5ndz AddedInterface Add eth0 [10.129.2.61/23] from ovn-kubernetes logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:07:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestallowedskunkmyjaeger-1-777866cbm5ndz.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:07:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestallowedskunkmyjaeger-1-777866cbm5ndz.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:07:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestallowedskunkmyjaeger-1-777866cbm5ndz.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:07:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestallowedskunkmyjaeger-1-777866cbm5ndz.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:07:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestallowedskunkmyjaeger-1-777866cbm5ndz.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:07:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestallowedskunkmyjaeger-1-777866cbm5ndz.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:07:55 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestallowedskunkmyjaeger-1-777866cbm5ndz.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:00 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestallowedskunkmyjaeger-1-777866cbm5ndz.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:11 +0000 UTC Normal Pod my-jaeger-collector-7c59c9c8fb-hj648 Binding Scheduled Successfully assigned kuttl-test-allowed-skunk/my-jaeger-collector-7c59c9c8fb-hj648 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:11 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7c59c9c8fb SuccessfulCreate Created pod: my-jaeger-collector-7c59c9c8fb-hj648 replicaset-controller logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:11 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7c59c9c8fb to 1 deployment-controller logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:11 +0000 UTC Normal Pod my-jaeger-query-649766b556-sqtth Binding Scheduled Successfully assigned kuttl-test-allowed-skunk/my-jaeger-query-649766b556-sqtth to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:11 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-649766b556 SuccessfulCreate Created pod: my-jaeger-query-649766b556-sqtth replicaset-controller logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:11 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-649766b556 to 1 deployment-controller logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:12 +0000 UTC Normal Pod my-jaeger-collector-7c59c9c8fb-hj648 AddedInterface Add eth0 [10.128.2.70/23] from ovn-kubernetes logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:12 +0000 UTC Normal Pod my-jaeger-collector-7c59c9c8fb-hj648.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:12 +0000 UTC Normal Pod my-jaeger-collector-7c59c9c8fb-hj648.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:12 +0000 UTC Normal Pod my-jaeger-collector-7c59c9c8fb-hj648.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:12 +0000 UTC Normal Pod my-jaeger-query-649766b556-sqtth AddedInterface Add eth0 [10.131.0.52/23] from ovn-kubernetes logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:12 +0000 UTC Normal Pod my-jaeger-query-649766b556-sqtth.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:12 +0000 UTC Normal Pod my-jaeger-query-649766b556-sqtth.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:12 +0000 UTC Normal Pod my-jaeger-query-649766b556-sqtth.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:12 +0000 UTC Normal Pod my-jaeger-query-649766b556-sqtth.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:12 +0000 UTC Normal Pod my-jaeger-query-649766b556-sqtth.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:12 +0000 UTC Normal Pod my-jaeger-query-649766b556-sqtth.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:12 +0000 UTC Normal Pod my-jaeger-query-649766b556-sqtth.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:12 +0000 UTC Normal Pod my-jaeger-query-649766b556-sqtth.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:12 +0000 UTC Normal Pod my-jaeger-query-649766b556-sqtth.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:17 +0000 UTC Normal Pod my-jaeger-query-649766b556-sqtth.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:17 +0000 UTC Normal Pod my-jaeger-query-649766b556-sqtth.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:17 +0000 UTC Normal Pod my-jaeger-query-649766b556-sqtth.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:17 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-649766b556 SuccessfulDelete Deleted pod: my-jaeger-query-649766b556-sqtth replicaset-controller logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:17 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-649766b556 to 0 from 1 deployment-controller logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:18 +0000 UTC Normal Pod my-jaeger-query-649848d6b7-plg44 Binding Scheduled Successfully assigned kuttl-test-allowed-skunk/my-jaeger-query-649848d6b7-plg44 to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:18 +0000 UTC Normal Pod my-jaeger-query-649848d6b7-plg44 AddedInterface Add eth0 [10.131.0.53/23] from ovn-kubernetes logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:18 +0000 UTC Normal Pod my-jaeger-query-649848d6b7-plg44.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:18 +0000 UTC Normal Pod my-jaeger-query-649848d6b7-plg44.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:18 +0000 UTC Normal Pod my-jaeger-query-649848d6b7-plg44.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:18 +0000 UTC Normal Pod my-jaeger-query-649848d6b7-plg44.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:18 +0000 UTC Normal Pod my-jaeger-query-649848d6b7-plg44.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:18 +0000 UTC Normal Pod my-jaeger-query-649848d6b7-plg44.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:18 +0000 UTC Normal Pod my-jaeger-query-649848d6b7-plg44.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:18 +0000 UTC Normal Pod my-jaeger-query-649848d6b7-plg44.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:18 +0000 UTC Normal Pod my-jaeger-query-649848d6b7-plg44.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:18 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-649848d6b7 SuccessfulCreate Created pod: my-jaeger-query-649848d6b7-plg44 replicaset-controller logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:18 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-649848d6b7 to 1 deployment-controller logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:23 +0000 UTC Normal Pod check-span-gzt2n Binding Scheduled Successfully assigned kuttl-test-allowed-skunk/check-span-gzt2n to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:23 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-gzt2n job-controller logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:23 +0000 UTC Normal Pod report-span-rjrk7 Binding Scheduled Successfully assigned kuttl-test-allowed-skunk/report-span-rjrk7 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:23 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-rjrk7 job-controller logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:24 +0000 UTC Normal Pod check-span-gzt2n AddedInterface Add eth0 [10.128.2.72/23] from ovn-kubernetes logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:24 +0000 UTC Normal Pod check-span-gzt2n.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:24 +0000 UTC Normal Pod check-span-gzt2n.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:24 +0000 UTC Normal Pod check-span-gzt2n.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:24 +0000 UTC Normal Pod report-span-rjrk7 AddedInterface Add eth0 [10.128.2.71/23] from ovn-kubernetes logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:24 +0000 UTC Normal Pod report-span-rjrk7.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:24 +0000 UTC Normal Pod report-span-rjrk7.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:24 +0000 UTC Normal Pod report-span-rjrk7.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:26 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:26 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:26 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:08:35 | collector-otlp-production-http | 2023-12-25 08:08:34 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:08:35 | collector-otlp-production-http | Deleting namespace: kuttl-test-allowed-skunk === CONT kuttl/harness/collector-otlp-allinone-grpc logger.go:42: 08:08:49 | collector-otlp-allinone-grpc | Creating namespace: kuttl-test-fitting-antelope logger.go:42: 08:08:49 | collector-otlp-allinone-grpc/0-install | starting test step 0-install logger.go:42: 08:08:49 | collector-otlp-allinone-grpc/0-install | Jaeger:kuttl-test-fitting-antelope/my-jaeger created logger.go:42: 08:08:54 | collector-otlp-allinone-grpc/0-install | test step completed 0-install logger.go:42: 08:08:54 | collector-otlp-allinone-grpc/1-smoke-test | starting test step 1-smoke-test logger.go:42: 08:08:54 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 08:08:55 | collector-otlp-allinone-grpc/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:09:02 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 08:09:02 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:09:03 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/report-span created logger.go:42: 08:09:03 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/check-span created logger.go:42: 08:09:22 | collector-otlp-allinone-grpc/1-smoke-test | test step completed 1-smoke-test logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | collector-otlp-allinone-grpc events from ns kuttl-test-fitting-antelope: logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:08:52 +0000 UTC Normal Pod my-jaeger-985ccd544-46hwl Binding Scheduled Successfully assigned kuttl-test-fitting-antelope/my-jaeger-985ccd544-46hwl to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:08:52 +0000 UTC Normal ReplicaSet.apps my-jaeger-985ccd544 SuccessfulCreate Created pod: my-jaeger-985ccd544-46hwl replicaset-controller logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:08:52 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-985ccd544 to 1 deployment-controller logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:08:53 +0000 UTC Normal Pod my-jaeger-985ccd544-46hwl AddedInterface Add eth0 [10.129.2.62/23] from ovn-kubernetes logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:08:53 +0000 UTC Normal Pod my-jaeger-985ccd544-46hwl.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:08:53 +0000 UTC Normal Pod my-jaeger-985ccd544-46hwl.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:08:53 +0000 UTC Normal Pod my-jaeger-985ccd544-46hwl.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:08:53 +0000 UTC Normal Pod my-jaeger-985ccd544-46hwl.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:08:53 +0000 UTC Normal Pod my-jaeger-985ccd544-46hwl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:08:53 +0000 UTC Normal Pod my-jaeger-985ccd544-46hwl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:08:59 +0000 UTC Normal Pod my-jaeger-985ccd544-46hwl.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:08:59 +0000 UTC Normal Pod my-jaeger-985ccd544-46hwl.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:08:59 +0000 UTC Normal ReplicaSet.apps my-jaeger-985ccd544 SuccessfulDelete Deleted pod: my-jaeger-985ccd544-46hwl replicaset-controller logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:08:59 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-985ccd544 to 0 from 1 deployment-controller logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:01 +0000 UTC Normal Pod my-jaeger-74c8d4d9c-c8m6m Binding Scheduled Successfully assigned kuttl-test-fitting-antelope/my-jaeger-74c8d4d9c-c8m6m to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:01 +0000 UTC Normal Pod my-jaeger-74c8d4d9c-c8m6m AddedInterface Add eth0 [10.129.2.63/23] from ovn-kubernetes logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:01 +0000 UTC Normal Pod my-jaeger-74c8d4d9c-c8m6m.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:01 +0000 UTC Normal Pod my-jaeger-74c8d4d9c-c8m6m.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:01 +0000 UTC Normal Pod my-jaeger-74c8d4d9c-c8m6m.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:01 +0000 UTC Normal Pod my-jaeger-74c8d4d9c-c8m6m.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:01 +0000 UTC Normal Pod my-jaeger-74c8d4d9c-c8m6m.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:01 +0000 UTC Normal Pod my-jaeger-74c8d4d9c-c8m6m.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:01 +0000 UTC Normal ReplicaSet.apps my-jaeger-74c8d4d9c SuccessfulCreate Created pod: my-jaeger-74c8d4d9c-c8m6m replicaset-controller logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:01 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-74c8d4d9c to 1 deployment-controller logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:03 +0000 UTC Normal Pod check-span-dtcnq Binding Scheduled Successfully assigned kuttl-test-fitting-antelope/check-span-dtcnq to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:03 +0000 UTC Normal Pod check-span-dtcnq AddedInterface Add eth0 [10.131.0.54/23] from ovn-kubernetes logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:03 +0000 UTC Normal Pod check-span-dtcnq.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:03 +0000 UTC Normal Pod check-span-dtcnq.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:03 +0000 UTC Normal Pod check-span-dtcnq.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:03 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-dtcnq job-controller logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:03 +0000 UTC Normal Pod report-span-q6mlb Binding Scheduled Successfully assigned kuttl-test-fitting-antelope/report-span-q6mlb to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:03 +0000 UTC Normal Pod report-span-q6mlb AddedInterface Add eth0 [10.128.2.73/23] from ovn-kubernetes logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:03 +0000 UTC Normal Pod report-span-q6mlb.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:03 +0000 UTC Normal Pod report-span-q6mlb.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:03 +0000 UTC Normal Pod report-span-q6mlb.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:03 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-q6mlb job-controller logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | 2023-12-25 08:09:22 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:09:22 | collector-otlp-allinone-grpc | Deleting namespace: kuttl-test-fitting-antelope === CONT kuttl/harness/collector-otlp-allinone-http logger.go:42: 08:09:35 | collector-otlp-allinone-http | Creating namespace: kuttl-test-creative-bison logger.go:42: 08:09:35 | collector-otlp-allinone-http/0-install | starting test step 0-install logger.go:42: 08:09:35 | collector-otlp-allinone-http/0-install | Jaeger:kuttl-test-creative-bison/my-jaeger created logger.go:42: 08:09:41 | collector-otlp-allinone-http/0-install | test step completed 0-install logger.go:42: 08:09:41 | collector-otlp-allinone-http/1-smoke-test | starting test step 1-smoke-test logger.go:42: 08:09:41 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 08:09:43 | collector-otlp-allinone-http/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:09:49 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7 OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 08:09:49 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:09:50 | collector-otlp-allinone-http/1-smoke-test | job.batch/report-span created logger.go:42: 08:09:50 | collector-otlp-allinone-http/1-smoke-test | job.batch/check-span created logger.go:42: 08:10:01 | collector-otlp-allinone-http/1-smoke-test | test step completed 1-smoke-test logger.go:42: 08:10:01 | collector-otlp-allinone-http | collector-otlp-allinone-http events from ns kuttl-test-creative-bison: logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:39 +0000 UTC Normal Pod my-jaeger-5db64bd6f4-mstvx Binding Scheduled Successfully assigned kuttl-test-creative-bison/my-jaeger-5db64bd6f4-mstvx to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:39 +0000 UTC Normal Pod my-jaeger-5db64bd6f4-mstvx AddedInterface Add eth0 [10.129.2.64/23] from ovn-kubernetes logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:39 +0000 UTC Normal Pod my-jaeger-5db64bd6f4-mstvx.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:39 +0000 UTC Normal Pod my-jaeger-5db64bd6f4-mstvx.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:39 +0000 UTC Normal Pod my-jaeger-5db64bd6f4-mstvx.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:39 +0000 UTC Normal Pod my-jaeger-5db64bd6f4-mstvx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:39 +0000 UTC Normal Pod my-jaeger-5db64bd6f4-mstvx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:39 +0000 UTC Normal Pod my-jaeger-5db64bd6f4-mstvx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:39 +0000 UTC Normal ReplicaSet.apps my-jaeger-5db64bd6f4 SuccessfulCreate Created pod: my-jaeger-5db64bd6f4-mstvx replicaset-controller logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:39 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5db64bd6f4 to 1 deployment-controller logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:46 +0000 UTC Normal Pod my-jaeger-5db64bd6f4-mstvx.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:46 +0000 UTC Normal Pod my-jaeger-5db64bd6f4-mstvx.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:46 +0000 UTC Normal ReplicaSet.apps my-jaeger-5db64bd6f4 SuccessfulDelete Deleted pod: my-jaeger-5db64bd6f4-mstvx replicaset-controller logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:46 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-5db64bd6f4 to 0 from 1 deployment-controller logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:47 +0000 UTC Normal Pod my-jaeger-c64f67f54-z4rv6 Binding Scheduled Successfully assigned kuttl-test-creative-bison/my-jaeger-c64f67f54-z4rv6 to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:47 +0000 UTC Normal Pod my-jaeger-c64f67f54-z4rv6 AddedInterface Add eth0 [10.129.2.65/23] from ovn-kubernetes logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:47 +0000 UTC Normal Pod my-jaeger-c64f67f54-z4rv6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:47 +0000 UTC Normal Pod my-jaeger-c64f67f54-z4rv6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:47 +0000 UTC Normal Pod my-jaeger-c64f67f54-z4rv6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:47 +0000 UTC Normal Pod my-jaeger-c64f67f54-z4rv6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:47 +0000 UTC Normal Pod my-jaeger-c64f67f54-z4rv6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:47 +0000 UTC Normal Pod my-jaeger-c64f67f54-z4rv6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:47 +0000 UTC Normal ReplicaSet.apps my-jaeger-c64f67f54 SuccessfulCreate Created pod: my-jaeger-c64f67f54-z4rv6 replicaset-controller logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:47 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-c64f67f54 to 1 deployment-controller logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:50 +0000 UTC Normal Pod check-span-sx56b Binding Scheduled Successfully assigned kuttl-test-creative-bison/check-span-sx56b to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:50 +0000 UTC Normal Pod check-span-sx56b AddedInterface Add eth0 [10.131.0.55/23] from ovn-kubernetes logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:50 +0000 UTC Normal Pod check-span-sx56b.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:50 +0000 UTC Normal Pod check-span-sx56b.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:50 +0000 UTC Normal Pod check-span-sx56b.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:50 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-sx56b job-controller logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:50 +0000 UTC Normal Pod report-span-bk89g Binding Scheduled Successfully assigned kuttl-test-creative-bison/report-span-bk89g to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:50 +0000 UTC Normal Pod report-span-bk89g AddedInterface Add eth0 [10.128.2.74/23] from ovn-kubernetes logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:50 +0000 UTC Normal Pod report-span-bk89g.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:50 +0000 UTC Normal Pod report-span-bk89g.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:50 +0000 UTC Normal Pod report-span-bk89g.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:09:50 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-bk89g job-controller logger.go:42: 08:10:01 | collector-otlp-allinone-http | 2023-12-25 08:10:01 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:10:01 | collector-otlp-allinone-http | Deleting namespace: kuttl-test-creative-bison === CONT kuttl/harness/collector-autoscale logger.go:42: 08:10:14 | collector-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:10:14 | collector-autoscale | Creating namespace: kuttl-test-allowed-bulldog logger.go:42: 08:10:14 | collector-autoscale/1-install | starting test step 1-install logger.go:42: 08:10:14 | collector-autoscale/1-install | Jaeger:kuttl-test-allowed-bulldog/simple-prod created logger.go:42: 08:10:52 | collector-autoscale/1-install | test step completed 1-install logger.go:42: 08:10:52 | collector-autoscale/2- | starting test step 2- logger.go:42: 08:10:52 | collector-autoscale/2- | test step completed 2- logger.go:42: 08:10:52 | collector-autoscale | collector-autoscale events from ns kuttl-test-allowed-bulldog: logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:21 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestallowedbulldogsimpleprod-1-6cdf47b9 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestallowedbulldogsimpleprod-1-6cdfmdf7h replicaset-controller logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestallowedbulldogsimpleprod-1-6cdfmdf7h Binding Scheduled Successfully assigned kuttl-test-allowed-bulldog/elasticsearch-cdm-kuttltestallowedbulldogsimpleprod-1-6cdfmdf7h to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:21 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestallowedbulldogsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestallowedbulldogsimpleprod-1-6cdf47b9 to 1 deployment-controller logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestallowedbulldogsimpleprod-1-6cdfmdf7h AddedInterface Add eth0 [10.129.2.66/23] from ovn-kubernetes logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestallowedbulldogsimpleprod-1-6cdfmdf7h.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestallowedbulldogsimpleprod-1-6cdfmdf7h.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestallowedbulldogsimpleprod-1-6cdfmdf7h.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestallowedbulldogsimpleprod-1-6cdfmdf7h.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestallowedbulldogsimpleprod-1-6cdfmdf7h.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestallowedbulldogsimpleprod-1-6cdfmdf7h.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:32 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestallowedbulldogsimpleprod-1-6cdfmdf7h.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:37 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestallowedbulldogsimpleprod-1-6cdfmdf7h.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:48 +0000 UTC Normal Pod simple-prod-collector-745d89545f-sp9dd Binding Scheduled Successfully assigned kuttl-test-allowed-bulldog/simple-prod-collector-745d89545f-sp9dd to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:48 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-745d89545f SuccessfulCreate Created pod: simple-prod-collector-745d89545f-sp9dd replicaset-controller logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:48 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-745d89545f to 1 deployment-controller logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:48 +0000 UTC Normal Pod simple-prod-query-74dbd878dc-qqgqm Binding Scheduled Successfully assigned kuttl-test-allowed-bulldog/simple-prod-query-74dbd878dc-qqgqm to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:48 +0000 UTC Normal ReplicaSet.apps simple-prod-query-74dbd878dc SuccessfulCreate Created pod: simple-prod-query-74dbd878dc-qqgqm replicaset-controller logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:48 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-74dbd878dc to 1 deployment-controller logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:49 +0000 UTC Normal Pod simple-prod-collector-745d89545f-sp9dd AddedInterface Add eth0 [10.128.2.75/23] from ovn-kubernetes logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:49 +0000 UTC Normal Pod simple-prod-collector-745d89545f-sp9dd.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:49 +0000 UTC Normal Pod simple-prod-collector-745d89545f-sp9dd.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:49 +0000 UTC Normal Pod simple-prod-collector-745d89545f-sp9dd.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:49 +0000 UTC Warning Pod simple-prod-query-74dbd878dc-qqgqm FailedMount MountVolume.SetUp failed for volume "simple-prod-ui-oauth-proxy-tls" : secret "simple-prod-ui-oauth-proxy-tls" not found kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:49 +0000 UTC Normal Pod simple-prod-query-74dbd878dc-qqgqm AddedInterface Add eth0 [10.131.0.56/23] from ovn-kubernetes logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:49 +0000 UTC Normal Pod simple-prod-query-74dbd878dc-qqgqm.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:50 +0000 UTC Normal Pod simple-prod-query-74dbd878dc-qqgqm.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:50 +0000 UTC Normal Pod simple-prod-query-74dbd878dc-qqgqm.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:50 +0000 UTC Normal Pod simple-prod-query-74dbd878dc-qqgqm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:50 +0000 UTC Normal Pod simple-prod-query-74dbd878dc-qqgqm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:50 +0000 UTC Normal Pod simple-prod-query-74dbd878dc-qqgqm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:50 +0000 UTC Normal Pod simple-prod-query-74dbd878dc-qqgqm.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:50 +0000 UTC Normal Pod simple-prod-query-74dbd878dc-qqgqm.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:10:52 | collector-autoscale | 2023-12-25 08:10:50 +0000 UTC Normal Pod simple-prod-query-74dbd878dc-qqgqm.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:10:52 | collector-autoscale | Deleting namespace: kuttl-test-allowed-bulldog === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (329.87s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.22s) --- PASS: kuttl/harness/collector-otlp-production-grpc (71.33s) --- PASS: kuttl/harness/set-custom-img (50.43s) --- PASS: kuttl/harness/collector-otlp-production-http (71.36s) --- PASS: kuttl/harness/collector-otlp-allinone-grpc (46.24s) --- PASS: kuttl/harness/collector-otlp-allinone-http (38.89s) --- PASS: kuttl/harness/collector-autoscale (45.24s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name miscellaneous --report --output /logs/artifacts/miscellaneous.xml ./artifacts/kuttl-report.xml time="2023-12-25T08:10:59Z" level=debug msg="Setting a new name for the test suites" time="2023-12-25T08:10:59Z" level=debug msg="Removing 'artifacts' TestCase" time="2023-12-25T08:10:59Z" level=debug msg="normalizing test case names" time="2023-12-25T08:10:59Z" level=debug msg="miscellaneous/artifacts -> miscellaneous_artifacts" time="2023-12-25T08:10:59Z" level=debug msg="miscellaneous/collector-otlp-production-grpc -> miscellaneous_collector_otlp_production_grpc" time="2023-12-25T08:10:59Z" level=debug msg="miscellaneous/set-custom-img -> miscellaneous_set_custom_img" time="2023-12-25T08:10:59Z" level=debug msg="miscellaneous/collector-otlp-production-http -> miscellaneous_collector_otlp_production_http" time="2023-12-25T08:10:59Z" level=debug msg="miscellaneous/collector-otlp-allinone-grpc -> miscellaneous_collector_otlp_allinone_grpc" time="2023-12-25T08:10:59Z" level=debug msg="miscellaneous/collector-otlp-allinone-http -> miscellaneous_collector_otlp_allinone_http" time="2023-12-25T08:10:59Z" level=debug msg="miscellaneous/collector-autoscale -> miscellaneous_collector_autoscale" +----------------------------------------------+--------+ | NAME | RESULT | +----------------------------------------------+--------+ | miscellaneous_artifacts | passed | | miscellaneous_collector_otlp_production_grpc | passed | | miscellaneous_set_custom_img | passed | | miscellaneous_collector_otlp_production_http | passed | | miscellaneous_collector_otlp_allinone_grpc | passed | | miscellaneous_collector_otlp_allinone_http | passed | | miscellaneous_collector_autoscale | passed | +----------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh sidecar false true + '[' 3 -ne 3 ']' + test_suite_name=sidecar + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/sidecar.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-sidecar make[2]: Entering directory '/tmp/jaeger-tests' ./tests/e2e/sidecar/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-12-22-165425 True False 44m Cluster version is 4.15.0-0.nightly-2023-12-22-165425' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-12-22-165425 True False 44m Cluster version is 4.15.0-0.nightly-2023-12-22-165425' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/sidecar/render.sh ++ export SUITE_DIR=./tests/e2e/sidecar ++ SUITE_DIR=./tests/e2e/sidecar ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/sidecar ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + jaeger_service_name=order + start_test sidecar-deployment + '[' 1 -ne 1 ']' + test_name=sidecar-deployment + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-deployment' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-deployment\e[0m' Rendering files for test sidecar-deployment + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build + '[' _build '!=' _build ']' + mkdir -p sidecar-deployment + cd sidecar-deployment + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-namespace + '[' 1 -ne 1 ']' + test_name=sidecar-namespace + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-namespace' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-namespace\e[0m' Rendering files for test sidecar-namespace + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-deployment + '[' sidecar-deployment '!=' _build ']' + cd .. + mkdir -p sidecar-namespace + cd sidecar-namespace + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-skip-webhook + '[' 1 -ne 1 ']' + test_name=sidecar-skip-webhook + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-skip-webhook' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-skip-webhook\e[0m' Rendering files for test sidecar-skip-webhook + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-namespace + '[' sidecar-namespace '!=' _build ']' + cd .. + mkdir -p sidecar-skip-webhook + cd sidecar-skip-webhook + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running sidecar E2E tests' Running sidecar E2E tests + cd tests/e2e/sidecar/_build + set +e + KUBECONFIG=/tmp/kubeconfig-1441553849 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/sidecar-deployment === PAUSE kuttl/harness/sidecar-deployment === RUN kuttl/harness/sidecar-namespace === PAUSE kuttl/harness/sidecar-namespace === RUN kuttl/harness/sidecar-skip-webhook === PAUSE kuttl/harness/sidecar-skip-webhook === CONT kuttl/harness/artifacts logger.go:42: 08:11:07 | artifacts | Creating namespace: kuttl-test-darling-yeti logger.go:42: 08:11:07 | artifacts | artifacts events from ns kuttl-test-darling-yeti: logger.go:42: 08:11:07 | artifacts | Deleting namespace: kuttl-test-darling-yeti === CONT kuttl/harness/sidecar-namespace logger.go:42: 08:11:13 | sidecar-namespace | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:11:14 | sidecar-namespace | Creating namespace: kuttl-test-better-seagull logger.go:42: 08:11:14 | sidecar-namespace/0-install | starting test step 0-install logger.go:42: 08:11:14 | sidecar-namespace/0-install | Jaeger:kuttl-test-better-seagull/agent-as-sidecar created logger.go:42: 08:11:20 | sidecar-namespace/0-install | test step completed 0-install logger.go:42: 08:11:20 | sidecar-namespace/1-install | starting test step 1-install logger.go:42: 08:11:20 | sidecar-namespace/1-install | Deployment:kuttl-test-better-seagull/vertx-create-span-sidecar created logger.go:42: 08:11:25 | sidecar-namespace/1-install | test step completed 1-install logger.go:42: 08:11:25 | sidecar-namespace/2-enable-injection | starting test step 2-enable-injection logger.go:42: 08:11:25 | sidecar-namespace/2-enable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="true"] logger.go:42: 08:11:25 | sidecar-namespace/2-enable-injection | namespace/kuttl-test-better-seagull annotate logger.go:42: 08:11:26 | sidecar-namespace/2-enable-injection | test step completed 2-enable-injection logger.go:42: 08:11:26 | sidecar-namespace/3-find-service | starting test step 3-find-service logger.go:42: 08:11:27 | sidecar-namespace/3-find-service | Job:kuttl-test-better-seagull/00-find-service created logger.go:42: 08:11:38 | sidecar-namespace/3-find-service | test step completed 3-find-service logger.go:42: 08:11:38 | sidecar-namespace/4-other-instance | starting test step 4-other-instance logger.go:42: 08:11:38 | sidecar-namespace/4-other-instance | Jaeger:kuttl-test-better-seagull/agent-as-sidecar2 created logger.go:42: 08:11:47 | sidecar-namespace/4-other-instance | test step completed 4-other-instance logger.go:42: 08:11:47 | sidecar-namespace/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 08:11:47 | sidecar-namespace/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 08:11:47 | sidecar-namespace/6-find-service | starting test step 6-find-service logger.go:42: 08:11:47 | sidecar-namespace/6-find-service | Job:kuttl-test-better-seagull/01-find-service created logger.go:42: 08:12:07 | sidecar-namespace/6-find-service | test step completed 6-find-service logger.go:42: 08:12:07 | sidecar-namespace/7-disable-injection | starting test step 7-disable-injection logger.go:42: 08:12:07 | sidecar-namespace/7-disable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="false"] logger.go:42: 08:12:07 | sidecar-namespace/7-disable-injection | namespace/kuttl-test-better-seagull annotate logger.go:42: 08:12:09 | sidecar-namespace/7-disable-injection | test step completed 7-disable-injection logger.go:42: 08:12:09 | sidecar-namespace | sidecar-namespace events from ns kuttl-test-better-seagull: logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:17 +0000 UTC Normal Pod agent-as-sidecar-6c8758496c-9h24s Binding Scheduled Successfully assigned kuttl-test-better-seagull/agent-as-sidecar-6c8758496c-9h24s to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:17 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-6c8758496c SuccessfulCreate Created pod: agent-as-sidecar-6c8758496c-9h24s replicaset-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:17 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-6c8758496c to 1 deployment-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:18 +0000 UTC Normal Pod agent-as-sidecar-6c8758496c-9h24s AddedInterface Add eth0 [10.129.2.67/23] from ovn-kubernetes logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:18 +0000 UTC Normal Pod agent-as-sidecar-6c8758496c-9h24s.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:18 +0000 UTC Normal Pod agent-as-sidecar-6c8758496c-9h24s.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:18 +0000 UTC Normal Pod agent-as-sidecar-6c8758496c-9h24s.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:20 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-r2t7f Binding Scheduled Successfully assigned kuttl-test-better-seagull/vertx-create-span-sidecar-84d458b68c-r2t7f to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:20 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-r2t7f AddedInterface Add eth0 [10.128.2.76/23] from ovn-kubernetes logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:20 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-r2t7f.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:20 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-r2t7f replicaset-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:20 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:24 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-r2t7f.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.729s (3.729s including waiting) kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:24 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-r2t7f.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:24 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-r2t7f.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:25 +0000 UTC Normal Pod vertx-create-span-sidecar-74699f68c7-p5fbx Binding Scheduled Successfully assigned kuttl-test-better-seagull/vertx-create-span-sidecar-74699f68c7-p5fbx to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:25 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-74699f68c7 SuccessfulCreate Created pod: vertx-create-span-sidecar-74699f68c7-p5fbx replicaset-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:25 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-74699f68c7 to 1 deployment-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:26 +0000 UTC Normal Pod vertx-create-span-sidecar-74699f68c7-p5fbx AddedInterface Add eth0 [10.131.0.57/23] from ovn-kubernetes logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:26 +0000 UTC Normal Pod vertx-create-span-sidecar-74699f68c7-p5fbx.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:26 +0000 UTC Normal Pod vertx-create-span-sidecar-74699f68c7-p5fbx.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:26 +0000 UTC Normal Pod vertx-create-span-sidecar-74699f68c7-p5fbx.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:26 +0000 UTC Normal Pod vertx-create-span-sidecar-74699f68c7-p5fbx.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:26 +0000 UTC Normal Pod vertx-create-span-sidecar-74699f68c7-p5fbx.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:26 +0000 UTC Normal Pod vertx-create-span-sidecar-74699f68c7-p5fbx.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:27 +0000 UTC Normal Pod 00-find-service-gkvcc Binding Scheduled Successfully assigned kuttl-test-better-seagull/00-find-service-gkvcc to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:27 +0000 UTC Normal Pod 00-find-service-gkvcc AddedInterface Add eth0 [10.129.2.68/23] from ovn-kubernetes logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:27 +0000 UTC Normal Pod 00-find-service-gkvcc.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:27 +0000 UTC Normal Pod 00-find-service-gkvcc.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:27 +0000 UTC Normal Pod 00-find-service-gkvcc.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:27 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-gkvcc job-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:32 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-r2t7f.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.76:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:32 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-r2t7f.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.76:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:34 +0000 UTC Warning Pod vertx-create-span-sidecar-74699f68c7-p5fbx.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.57:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:34 +0000 UTC Warning Pod vertx-create-span-sidecar-74699f68c7-p5fbx.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.57:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:35 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-r2t7f.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:35 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-r2t7f.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.76:8080/": read tcp 10.128.2.2:58994->10.128.2.76:8080: read: connection reset by peer kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:35 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-r2t7f.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.76:8080/": dial tcp 10.128.2.76:8080: connect: connection refused kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:35 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-r2t7f.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:36 +0000 UTC Normal Pod vertx-create-span-sidecar-74699f68c7-p5fbx.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:36 +0000 UTC Warning Pod vertx-create-span-sidecar-74699f68c7-p5fbx.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.57:8080/": read tcp 10.131.0.2:42816->10.131.0.57:8080: read: connection reset by peer kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:36 +0000 UTC Warning Pod vertx-create-span-sidecar-74699f68c7-p5fbx.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.57:8080/": dial tcp 10.131.0.57:8080: connect: connection refused kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:37 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:44 +0000 UTC Normal Pod agent-as-sidecar2-54bd5cd77c-pmk27 Binding Scheduled Successfully assigned kuttl-test-better-seagull/agent-as-sidecar2-54bd5cd77c-pmk27 to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:44 +0000 UTC Normal Pod agent-as-sidecar2-54bd5cd77c-pmk27 AddedInterface Add eth0 [10.129.2.69/23] from ovn-kubernetes logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:44 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-54bd5cd77c SuccessfulCreate Created pod: agent-as-sidecar2-54bd5cd77c-pmk27 replicaset-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:44 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-54bd5cd77c to 1 deployment-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:45 +0000 UTC Normal Pod agent-as-sidecar2-54bd5cd77c-pmk27.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:45 +0000 UTC Normal Pod agent-as-sidecar2-54bd5cd77c-pmk27.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:45 +0000 UTC Normal Pod agent-as-sidecar2-54bd5cd77c-pmk27.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:46 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-r2t7f.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.76:8080/": read tcp 10.128.2.2:52644->10.128.2.76:8080: read: connection reset by peer kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:47 +0000 UTC Normal Pod 01-find-service-nxw2b Binding Scheduled Successfully assigned kuttl-test-better-seagull/01-find-service-nxw2b to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:47 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-nxw2b job-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:47 +0000 UTC Normal Pod agent-as-sidecar-6c8758496c-9h24s.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:48 +0000 UTC Normal Pod 01-find-service-nxw2b AddedInterface Add eth0 [10.129.2.70/23] from ovn-kubernetes logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:48 +0000 UTC Normal Pod 01-find-service-nxw2b.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:48 +0000 UTC Normal Pod 01-find-service-nxw2b.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:48 +0000 UTC Normal Pod 01-find-service-nxw2b.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:48 +0000 UTC Warning Pod agent-as-sidecar-6c8758496c-9h24s.spec.containers{jaeger} Unhealthy Readiness probe failed: Get "http://10.129.2.67:14269/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:51 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-r2t7f replicaset-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:51 +0000 UTC Normal Pod vertx-create-span-sidecar-8c6c7c959-r46qd Binding Scheduled Successfully assigned kuttl-test-better-seagull/vertx-create-span-sidecar-8c6c7c959-r46qd to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:51 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-8c6c7c959 SuccessfulCreate Created pod: vertx-create-span-sidecar-8c6c7c959-r46qd replicaset-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:51 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:51 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-8c6c7c959 to 1 from 0 deployment-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:52 +0000 UTC Normal Pod vertx-create-span-sidecar-8c6c7c959-r46qd AddedInterface Add eth0 [10.128.2.77/23] from ovn-kubernetes logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:52 +0000 UTC Normal Pod vertx-create-span-sidecar-8c6c7c959-r46qd.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:52 +0000 UTC Normal Pod vertx-create-span-sidecar-8c6c7c959-r46qd.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:52 +0000 UTC Normal Pod vertx-create-span-sidecar-8c6c7c959-r46qd.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:52 +0000 UTC Normal Pod vertx-create-span-sidecar-8c6c7c959-r46qd.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:52 +0000 UTC Normal Pod vertx-create-span-sidecar-8c6c7c959-r46qd.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:11:52 +0000 UTC Normal Pod vertx-create-span-sidecar-8c6c7c959-r46qd.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:12:00 +0000 UTC Warning Pod vertx-create-span-sidecar-8c6c7c959-r46qd.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.77:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:12:00 +0000 UTC Warning Pod vertx-create-span-sidecar-8c6c7c959-r46qd.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.77:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:12:02 +0000 UTC Normal Pod vertx-create-span-sidecar-8c6c7c959-r46qd.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:12:02 +0000 UTC Warning Pod vertx-create-span-sidecar-8c6c7c959-r46qd.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.77:8080/": read tcp 10.128.2.2:41464->10.128.2.77:8080: read: connection reset by peer kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:12:02 +0000 UTC Warning Pod vertx-create-span-sidecar-8c6c7c959-r46qd.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.77:8080/": dial tcp 10.128.2.77:8080: connect: connection refused kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:12:06 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:12:07 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-74699f68c7 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-74699f68c7-p5fbx replicaset-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:12:07 +0000 UTC Normal Pod vertx-create-span-sidecar-ffc77dcc4-ktpkm Binding Scheduled Successfully assigned kuttl-test-better-seagull/vertx-create-span-sidecar-ffc77dcc4-ktpkm to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:12:07 +0000 UTC Normal Pod vertx-create-span-sidecar-ffc77dcc4-ktpkm AddedInterface Add eth0 [10.129.2.71/23] from ovn-kubernetes logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:12:07 +0000 UTC Normal Pod vertx-create-span-sidecar-ffc77dcc4-ktpkm.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:12:07 +0000 UTC Normal Pod vertx-create-span-sidecar-ffc77dcc4-ktpkm.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:12:07 +0000 UTC Normal Pod vertx-create-span-sidecar-ffc77dcc4-ktpkm.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:12:07 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-ffc77dcc4 SuccessfulCreate Created pod: vertx-create-span-sidecar-ffc77dcc4-ktpkm replicaset-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:12:07 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-74699f68c7 to 0 from 1 deployment-controller logger.go:42: 08:12:09 | sidecar-namespace | 2023-12-25 08:12:07 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-ffc77dcc4 to 1 from 0 deployment-controller logger.go:42: 08:12:09 | sidecar-namespace | Deleting namespace: kuttl-test-better-seagull === CONT kuttl/harness/sidecar-skip-webhook logger.go:42: 08:12:17 | sidecar-skip-webhook | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:12:17 | sidecar-skip-webhook | Creating namespace: kuttl-test-grand-whale logger.go:42: 08:12:17 | sidecar-skip-webhook/0-install | starting test step 0-install logger.go:42: 08:12:17 | sidecar-skip-webhook/0-install | Jaeger:kuttl-test-grand-whale/agent-as-sidecar created logger.go:42: 08:12:23 | sidecar-skip-webhook/0-install | test step completed 0-install logger.go:42: 08:12:23 | sidecar-skip-webhook/1-install | starting test step 1-install logger.go:42: 08:12:23 | sidecar-skip-webhook/1-install | Deployment:kuttl-test-grand-whale/vertx-create-span-sidecar created logger.go:42: 08:12:25 | sidecar-skip-webhook/1-install | test step completed 1-install logger.go:42: 08:12:25 | sidecar-skip-webhook/2-add-anotation-and-label | starting test step 2-add-anotation-and-label logger.go:42: 08:12:25 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name=jaeger-operator --namespace kuttl-test-grand-whale] logger.go:42: 08:12:25 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar labeled logger.go:42: 08:12:25 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-grand-whale] logger.go:42: 08:12:25 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar annotate logger.go:42: 08:12:25 | sidecar-skip-webhook/2-add-anotation-and-label | test step completed 2-add-anotation-and-label logger.go:42: 08:12:25 | sidecar-skip-webhook/3-remove-label | starting test step 3-remove-label logger.go:42: 08:12:25 | sidecar-skip-webhook/3-remove-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name- --namespace kuttl-test-grand-whale] logger.go:42: 08:12:26 | sidecar-skip-webhook/3-remove-label | deployment.apps/vertx-create-span-sidecar unlabeled logger.go:42: 08:12:27 | sidecar-skip-webhook/3-remove-label | test step completed 3-remove-label logger.go:42: 08:12:27 | sidecar-skip-webhook | sidecar-skip-webhook events from ns kuttl-test-grand-whale: logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:20 +0000 UTC Normal Pod agent-as-sidecar-7d87f56945-mdrgc Binding Scheduled Successfully assigned kuttl-test-grand-whale/agent-as-sidecar-7d87f56945-mdrgc to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:20 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-7d87f56945 SuccessfulCreate Created pod: agent-as-sidecar-7d87f56945-mdrgc replicaset-controller logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:20 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-7d87f56945 to 1 deployment-controller logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:21 +0000 UTC Normal Pod agent-as-sidecar-7d87f56945-mdrgc AddedInterface Add eth0 [10.129.2.72/23] from ovn-kubernetes logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:21 +0000 UTC Normal Pod agent-as-sidecar-7d87f56945-mdrgc.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:21 +0000 UTC Normal Pod agent-as-sidecar-7d87f56945-mdrgc.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:21 +0000 UTC Normal Pod agent-as-sidecar-7d87f56945-mdrgc.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:23 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-gh6xs Binding Scheduled Successfully assigned kuttl-test-grand-whale/vertx-create-span-sidecar-84d458b68c-gh6xs to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:23 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-gh6xs AddedInterface Add eth0 [10.128.2.78/23] from ovn-kubernetes logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:23 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-gh6xs.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:23 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-gh6xs.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:23 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-gh6xs.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:23 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-gh6xs replicaset-controller logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:23 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:25 +0000 UTC Normal Pod vertx-create-span-sidecar-54f55dd49c-jmx78 Binding Scheduled Successfully assigned kuttl-test-grand-whale/vertx-create-span-sidecar-54f55dd49c-jmx78 to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:25 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-54f55dd49c SuccessfulCreate Created pod: vertx-create-span-sidecar-54f55dd49c-jmx78 replicaset-controller logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:25 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-54f55dd49c to 1 deployment-controller logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:26 +0000 UTC Normal Pod vertx-create-span-sidecar-54f55dd49c-jmx78 AddedInterface Add eth0 [10.131.0.58/23] from ovn-kubernetes logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:26 +0000 UTC Normal Pod vertx-create-span-sidecar-54f55dd49c-jmx78.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:26 +0000 UTC Normal Pod vertx-create-span-sidecar-54f55dd49c-jmx78.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:26 +0000 UTC Normal Pod vertx-create-span-sidecar-54f55dd49c-jmx78.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:26 +0000 UTC Normal Pod vertx-create-span-sidecar-54f55dd49c-jmx78.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:26 +0000 UTC Normal Pod vertx-create-span-sidecar-54f55dd49c-jmx78.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:12:27 | sidecar-skip-webhook | 2023-12-25 08:12:26 +0000 UTC Normal Pod vertx-create-span-sidecar-54f55dd49c-jmx78.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:12:27 | sidecar-skip-webhook | Deleting namespace: kuttl-test-grand-whale === CONT kuttl/harness/sidecar-deployment logger.go:42: 08:12:33 | sidecar-deployment | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:12:33 | sidecar-deployment | Creating namespace: kuttl-test-sensible-shiner logger.go:42: 08:12:33 | sidecar-deployment/0-install | starting test step 0-install logger.go:42: 08:12:33 | sidecar-deployment/0-install | Jaeger:kuttl-test-sensible-shiner/agent-as-sidecar created logger.go:42: 08:12:40 | sidecar-deployment/0-install | test step completed 0-install logger.go:42: 08:12:40 | sidecar-deployment/1-install | starting test step 1-install logger.go:42: 08:12:40 | sidecar-deployment/1-install | Deployment:kuttl-test-sensible-shiner/vertx-create-span-sidecar created logger.go:42: 08:12:42 | sidecar-deployment/1-install | test step completed 1-install logger.go:42: 08:12:42 | sidecar-deployment/2-enable-injection | starting test step 2-enable-injection logger.go:42: 08:12:42 | sidecar-deployment/2-enable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-sensible-shiner] logger.go:42: 08:12:42 | sidecar-deployment/2-enable-injection | deployment.apps/vertx-create-span-sidecar annotate logger.go:42: 08:12:44 | sidecar-deployment/2-enable-injection | test step completed 2-enable-injection logger.go:42: 08:12:44 | sidecar-deployment/3-find-service | starting test step 3-find-service logger.go:42: 08:12:44 | sidecar-deployment/3-find-service | Job:kuttl-test-sensible-shiner/00-find-service created logger.go:42: 08:12:56 | sidecar-deployment/3-find-service | test step completed 3-find-service logger.go:42: 08:12:56 | sidecar-deployment/4-other-instance | starting test step 4-other-instance logger.go:42: 08:12:56 | sidecar-deployment/4-other-instance | Jaeger:kuttl-test-sensible-shiner/agent-as-sidecar2 created logger.go:42: 08:13:07 | sidecar-deployment/4-other-instance | test step completed 4-other-instance logger.go:42: 08:13:07 | sidecar-deployment/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 08:13:07 | sidecar-deployment/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 08:13:07 | sidecar-deployment/6-find-service | starting test step 6-find-service logger.go:42: 08:13:07 | sidecar-deployment/6-find-service | Job:kuttl-test-sensible-shiner/01-find-service created logger.go:42: 08:13:27 | sidecar-deployment/6-find-service | test step completed 6-find-service logger.go:42: 08:13:27 | sidecar-deployment/7-disable-injection | starting test step 7-disable-injection logger.go:42: 08:13:27 | sidecar-deployment/7-disable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=false --namespace kuttl-test-sensible-shiner] logger.go:42: 08:13:27 | sidecar-deployment/7-disable-injection | deployment.apps/vertx-create-span-sidecar annotate logger.go:42: 08:13:28 | sidecar-deployment/7-disable-injection | test step completed 7-disable-injection logger.go:42: 08:13:28 | sidecar-deployment | sidecar-deployment events from ns kuttl-test-sensible-shiner: logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:37 +0000 UTC Normal Pod agent-as-sidecar-59f4b78bd8-7k4cx Binding Scheduled Successfully assigned kuttl-test-sensible-shiner/agent-as-sidecar-59f4b78bd8-7k4cx to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:37 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-59f4b78bd8 SuccessfulCreate Created pod: agent-as-sidecar-59f4b78bd8-7k4cx replicaset-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:37 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-59f4b78bd8 to 1 deployment-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:38 +0000 UTC Normal Pod agent-as-sidecar-59f4b78bd8-7k4cx AddedInterface Add eth0 [10.129.2.73/23] from ovn-kubernetes logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:38 +0000 UTC Normal Pod agent-as-sidecar-59f4b78bd8-7k4cx.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:38 +0000 UTC Normal Pod agent-as-sidecar-59f4b78bd8-7k4cx.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:38 +0000 UTC Normal Pod agent-as-sidecar-59f4b78bd8-7k4cx.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:40 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-92k8v Binding Scheduled Successfully assigned kuttl-test-sensible-shiner/vertx-create-span-sidecar-84d458b68c-92k8v to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:40 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-92k8v AddedInterface Add eth0 [10.128.2.79/23] from ovn-kubernetes logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:40 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-92k8v replicaset-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:40 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:41 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-92k8v.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:41 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-92k8v.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:41 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-92k8v.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:42 +0000 UTC Normal Pod vertx-create-span-sidecar-684fddd6c4-wbl98 Binding Scheduled Successfully assigned kuttl-test-sensible-shiner/vertx-create-span-sidecar-684fddd6c4-wbl98 to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:42 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-684fddd6c4 SuccessfulCreate Created pod: vertx-create-span-sidecar-684fddd6c4-wbl98 replicaset-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:42 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-684fddd6c4 to 1 deployment-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:43 +0000 UTC Normal Pod vertx-create-span-sidecar-684fddd6c4-wbl98 AddedInterface Add eth0 [10.129.2.74/23] from ovn-kubernetes logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:43 +0000 UTC Normal Pod vertx-create-span-sidecar-684fddd6c4-wbl98.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:43 +0000 UTC Normal Pod vertx-create-span-sidecar-684fddd6c4-wbl98.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:43 +0000 UTC Normal Pod vertx-create-span-sidecar-684fddd6c4-wbl98.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:43 +0000 UTC Normal Pod vertx-create-span-sidecar-684fddd6c4-wbl98.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:43 +0000 UTC Normal Pod vertx-create-span-sidecar-684fddd6c4-wbl98.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:43 +0000 UTC Normal Pod vertx-create-span-sidecar-684fddd6c4-wbl98.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:44 +0000 UTC Normal Pod 00-find-service-55sf9 Binding Scheduled Successfully assigned kuttl-test-sensible-shiner/00-find-service-55sf9 to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:44 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-55sf9 job-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:45 +0000 UTC Normal Pod 00-find-service-55sf9 AddedInterface Add eth0 [10.131.0.59/23] from ovn-kubernetes logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:45 +0000 UTC Normal Pod 00-find-service-55sf9.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:45 +0000 UTC Normal Pod 00-find-service-55sf9.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:45 +0000 UTC Normal Pod 00-find-service-55sf9.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:49 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-92k8v.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.79:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:49 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-92k8v.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.79:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:51 +0000 UTC Warning Pod vertx-create-span-sidecar-684fddd6c4-wbl98.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.74:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:51 +0000 UTC Warning Pod vertx-create-span-sidecar-684fddd6c4-wbl98.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.74:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:52 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-92k8v.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:52 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-92k8v.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.79:8080/": read tcp 10.128.2.2:52072->10.128.2.79:8080: read: connection reset by peer kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:52 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-92k8v.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.79:8080/": dial tcp 10.128.2.79:8080: connect: connection refused kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:53 +0000 UTC Normal Pod vertx-create-span-sidecar-684fddd6c4-wbl98.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:53 +0000 UTC Warning Pod vertx-create-span-sidecar-684fddd6c4-wbl98.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.74:8080/": read tcp 10.129.2.2:56452->10.129.2.74:8080: read: connection reset by peer kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:53 +0000 UTC Warning Pod vertx-create-span-sidecar-684fddd6c4-wbl98.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.74:8080/": dial tcp 10.129.2.74:8080: connect: connection refused kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:55 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:59 +0000 UTC Normal Pod agent-as-sidecar2-6fbb7b97c6-7pnrl Binding Scheduled Successfully assigned kuttl-test-sensible-shiner/agent-as-sidecar2-6fbb7b97c6-7pnrl to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:59 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-6fbb7b97c6 SuccessfulCreate Created pod: agent-as-sidecar2-6fbb7b97c6-7pnrl replicaset-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:12:59 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-6fbb7b97c6 to 1 deployment-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:00 +0000 UTC Warning Pod agent-as-sidecar2-6fbb7b97c6-7pnrl FailedMount MountVolume.SetUp failed for volume "agent-as-sidecar2-collector-tls-config-volume" : secret "agent-as-sidecar2-collector-headless-tls" not found kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:00 +0000 UTC Normal Pod agent-as-sidecar2-6fbb7b97c6-7pnrl AddedInterface Add eth0 [10.131.0.60/23] from ovn-kubernetes logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:01 +0000 UTC Normal Pod agent-as-sidecar2-6fbb7b97c6-7pnrl.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:03 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-92k8v.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.79:8080/": read tcp 10.128.2.2:56028->10.128.2.79:8080: read: connection reset by peer kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:05 +0000 UTC Normal Pod agent-as-sidecar2-6fbb7b97c6-7pnrl.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" in 4.269s (4.269s including waiting) kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:05 +0000 UTC Normal Pod agent-as-sidecar2-6fbb7b97c6-7pnrl.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:05 +0000 UTC Normal Pod agent-as-sidecar2-6fbb7b97c6-7pnrl.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:07 +0000 UTC Normal Pod 01-find-service-6lx8f Binding Scheduled Successfully assigned kuttl-test-sensible-shiner/01-find-service-6lx8f to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:07 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-6lx8f job-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:07 +0000 UTC Normal Pod agent-as-sidecar-59f4b78bd8-7k4cx.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:08 +0000 UTC Normal Pod 01-find-service-6lx8f AddedInterface Add eth0 [10.131.0.61/23] from ovn-kubernetes logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:08 +0000 UTC Normal Pod 01-find-service-6lx8f.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-w25v5qjw/pipeline@sha256:c7719edf50d5c28790edb2f6f0c49aa938e718fda00df65c5c6bd3d17dd6a8c7" already present on machine kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:08 +0000 UTC Normal Pod 01-find-service-6lx8f.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:08 +0000 UTC Normal Pod 01-find-service-6lx8f.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:11 +0000 UTC Normal Pod vertx-create-span-sidecar-65dd75dcbd-th2xv Binding Scheduled Successfully assigned kuttl-test-sensible-shiner/vertx-create-span-sidecar-65dd75dcbd-th2xv to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:11 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-65dd75dcbd SuccessfulCreate Created pod: vertx-create-span-sidecar-65dd75dcbd-th2xv replicaset-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:11 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-92k8v replicaset-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:11 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:11 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-65dd75dcbd to 1 from 0 deployment-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:12 +0000 UTC Normal Pod vertx-create-span-sidecar-65dd75dcbd-th2xv AddedInterface Add eth0 [10.128.2.80/23] from ovn-kubernetes logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:12 +0000 UTC Normal Pod vertx-create-span-sidecar-65dd75dcbd-th2xv.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:12 +0000 UTC Normal Pod vertx-create-span-sidecar-65dd75dcbd-th2xv.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:12 +0000 UTC Normal Pod vertx-create-span-sidecar-65dd75dcbd-th2xv.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:12 +0000 UTC Normal Pod vertx-create-span-sidecar-65dd75dcbd-th2xv.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:12 +0000 UTC Normal Pod vertx-create-span-sidecar-65dd75dcbd-th2xv.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:12 +0000 UTC Normal Pod vertx-create-span-sidecar-65dd75dcbd-th2xv.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:20 +0000 UTC Warning Pod vertx-create-span-sidecar-65dd75dcbd-th2xv.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.80:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:20 +0000 UTC Warning Pod vertx-create-span-sidecar-65dd75dcbd-th2xv.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.80:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:22 +0000 UTC Normal Pod vertx-create-span-sidecar-65dd75dcbd-th2xv.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:22 +0000 UTC Warning Pod vertx-create-span-sidecar-65dd75dcbd-th2xv.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.80:8080/": read tcp 10.128.2.2:56608->10.128.2.80:8080: read: connection reset by peer kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:22 +0000 UTC Warning Pod vertx-create-span-sidecar-65dd75dcbd-th2xv.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.80:8080/": dial tcp 10.128.2.80:8080: connect: connection refused kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:27 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:27 +0000 UTC Normal Pod vertx-create-span-sidecar-66f667b97f-4hkh8 Binding Scheduled Successfully assigned kuttl-test-sensible-shiner/vertx-create-span-sidecar-66f667b97f-4hkh8 to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:27 +0000 UTC Normal Pod vertx-create-span-sidecar-66f667b97f-4hkh8 AddedInterface Add eth0 [10.131.0.62/23] from ovn-kubernetes logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:27 +0000 UTC Normal Pod vertx-create-span-sidecar-66f667b97f-4hkh8.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:27 +0000 UTC Normal Pod vertx-create-span-sidecar-66f667b97f-4hkh8.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:27 +0000 UTC Normal Pod vertx-create-span-sidecar-66f667b97f-4hkh8.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:27 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-66f667b97f SuccessfulCreate Created pod: vertx-create-span-sidecar-66f667b97f-4hkh8 replicaset-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:27 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-684fddd6c4 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-684fddd6c4-wbl98 replicaset-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:27 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-684fddd6c4 to 0 from 1 deployment-controller logger.go:42: 08:13:28 | sidecar-deployment | 2023-12-25 08:13:27 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-66f667b97f to 1 from 0 deployment-controller logger.go:42: 08:13:28 | sidecar-deployment | Deleting namespace: kuttl-test-sensible-shiner === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (148.19s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.24s) --- PASS: kuttl/harness/sidecar-namespace (63.06s) --- PASS: kuttl/harness/sidecar-skip-webhook (16.73s) --- PASS: kuttl/harness/sidecar-deployment (61.98s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name sidecar --report --output /logs/artifacts/sidecar.xml ./artifacts/kuttl-report.xml time="2023-12-25T08:13:36Z" level=debug msg="Setting a new name for the test suites" time="2023-12-25T08:13:36Z" level=debug msg="Removing 'artifacts' TestCase" time="2023-12-25T08:13:36Z" level=debug msg="normalizing test case names" time="2023-12-25T08:13:36Z" level=debug msg="sidecar/artifacts -> sidecar_artifacts" time="2023-12-25T08:13:36Z" level=debug msg="sidecar/sidecar-namespace -> sidecar_sidecar_namespace" time="2023-12-25T08:13:36Z" level=debug msg="sidecar/sidecar-skip-webhook -> sidecar_sidecar_skip_webhook" time="2023-12-25T08:13:36Z" level=debug msg="sidecar/sidecar-deployment -> sidecar_sidecar_deployment" +------------------------------+--------+ | NAME | RESULT | +------------------------------+--------+ | sidecar_artifacts | passed | | sidecar_sidecar_namespace | passed | | sidecar_sidecar_skip_webhook | passed | | sidecar_sidecar_deployment | passed | +------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh streaming false true + '[' 3 -ne 3 ']' + test_suite_name=streaming + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/streaming.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-streaming make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ SKIP_ES_EXTERNAL=true \ ./tests/e2e/streaming/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-12-22-165425 True False 47m Cluster version is 4.15.0-0.nightly-2023-12-22-165425' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-12-22-165425 True False 47m Cluster version is 4.15.0-0.nightly-2023-12-22-165425' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/streaming/render.sh ++ export SUITE_DIR=./tests/e2e/streaming ++ SUITE_DIR=./tests/e2e/streaming ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/streaming ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + '[' false = true ']' + start_test streaming-simple + '[' 1 -ne 1 ']' + test_name=streaming-simple + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-simple' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-simple\e[0m' Rendering files for test streaming-simple + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + mkdir -p streaming-simple + cd streaming-simple + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/streaming-jaeger-assert.yaml.template -o ./04-assert.yaml + render_smoke_test simple-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=simple-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + export JAEGER_NAME=simple-streaming + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-tls + '[' 1 -ne 1 ']' + test_name=streaming-with-tls + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-tls' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-tls\e[0m' Rendering files for test streaming-with-tls + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-simple + '[' streaming-simple '!=' _build ']' + cd .. + mkdir -p streaming-with-tls + cd streaming-with-tls + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + render_smoke_test tls-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=tls-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + export JAEGER_NAME=tls-streaming + JAEGER_NAME=tls-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-autoprovisioning-autoscale + '[' 1 -ne 1 ']' + test_name=streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-autoprovisioning-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-autoprovisioning-autoscale\e[0m' Rendering files for test streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-tls + '[' streaming-with-tls '!=' _build ']' + cd .. + mkdir -p streaming-with-autoprovisioning-autoscale + cd streaming-with-autoprovisioning-autoscale + '[' true = true ']' + rm ./00-install.yaml ./00-assert.yaml + render_install_elasticsearch upstream 01 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=01 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./01-assert.yaml + jaeger_name=auto-provisioned + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="20Mi"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="500m"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.autoscale=true ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.minReplicas=1 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.maxReplicas=2 ./02-install.yaml + render_assert_kafka true auto-provisioned 03 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provisioned + test_step=03 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./03-assert.yaml ++ expr 03 + 1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./04-assert.yaml ++ expr 03 + 2 + CLUSTER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./05-assert.yaml + version_lt 1.27 1.23 ++ echo 1.27 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.27 '!=' 1.27 + rm ./08-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running streaming E2E tests' Running streaming E2E tests + cd tests/e2e/streaming/_build + set +e + KUBECONFIG=/tmp/kubeconfig-1441553849 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/streaming-simple === PAUSE kuttl/harness/streaming-simple === RUN kuttl/harness/streaming-with-autoprovisioning-autoscale === PAUSE kuttl/harness/streaming-with-autoprovisioning-autoscale === RUN kuttl/harness/streaming-with-tls === PAUSE kuttl/harness/streaming-with-tls === CONT kuttl/harness/artifacts logger.go:42: 08:13:48 | artifacts | Creating namespace: kuttl-test-healthy-lamprey logger.go:42: 08:13:48 | artifacts | artifacts events from ns kuttl-test-healthy-lamprey: logger.go:42: 08:13:48 | artifacts | Deleting namespace: kuttl-test-healthy-lamprey === CONT kuttl/harness/streaming-with-autoprovisioning-autoscale logger.go:42: 08:13:54 | streaming-with-autoprovisioning-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:13:54 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:13:54 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:13:54 | streaming-with-autoprovisioning-autoscale | Creating namespace: kuttl-test-known-ray logger.go:42: 08:13:54 | streaming-with-autoprovisioning-autoscale/1-install | starting test step 1-install logger.go:42: 08:13:54 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc create sa deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 08:13:55 | streaming-with-autoprovisioning-autoscale/1-install | serviceaccount/deploy-elasticsearch created logger.go:42: 08:13:55 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc adm policy add-scc-to-user privileged -z deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 08:13:55 | streaming-with-autoprovisioning-autoscale/1-install | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:privileged added: "deploy-elasticsearch" logger.go:42: 08:13:55 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 6] logger.go:42: 08:14:01 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_0.yml -n $NAMESPACE] logger.go:42: 08:14:01 | streaming-with-autoprovisioning-autoscale/1-install | statefulset.apps/elasticsearch created logger.go:42: 08:14:01 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 3] logger.go:42: 08:14:04 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_1.yml -n $NAMESPACE] logger.go:42: 08:14:05 | streaming-with-autoprovisioning-autoscale/1-install | service/elasticsearch created logger.go:42: 08:14:22 | streaming-with-autoprovisioning-autoscale/1-install | test step completed 1-install logger.go:42: 08:14:22 | streaming-with-autoprovisioning-autoscale/2-install | starting test step 2-install logger.go:42: 08:14:22 | streaming-with-autoprovisioning-autoscale/2-install | Jaeger:kuttl-test-known-ray/auto-provisioned created logger.go:42: 08:14:22 | streaming-with-autoprovisioning-autoscale/2-install | test step completed 2-install logger.go:42: 08:14:22 | streaming-with-autoprovisioning-autoscale/3- | starting test step 3- logger.go:42: 08:15:10 | streaming-with-autoprovisioning-autoscale/3- | test step completed 3- logger.go:42: 08:15:10 | streaming-with-autoprovisioning-autoscale/4- | starting test step 4- logger.go:42: 08:15:41 | streaming-with-autoprovisioning-autoscale/4- | test step completed 4- logger.go:42: 08:15:41 | streaming-with-autoprovisioning-autoscale/5- | starting test step 5- logger.go:42: 08:16:13 | streaming-with-autoprovisioning-autoscale/5- | test step completed 5- logger.go:42: 08:16:13 | streaming-with-autoprovisioning-autoscale/6- | starting test step 6- logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale/6- | test step completed 6- logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale/7- | starting test step 7- logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale/7- | test step completed 7- logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | streaming-with-autoprovisioning-autoscale events from ns kuttl-test-known-ray: logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:01 +0000 UTC Normal Pod elasticsearch-0 Binding Scheduled Successfully assigned kuttl-test-known-ray/elasticsearch-0 to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:01 +0000 UTC Normal StatefulSet.apps elasticsearch SuccessfulCreate create Pod elasticsearch-0 in StatefulSet elasticsearch successful statefulset-controller logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:02 +0000 UTC Normal Pod elasticsearch-0 AddedInterface Add eth0 [10.129.2.75/23] from ovn-kubernetes logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:02 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulling Pulling image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:10 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulled Successfully pulled image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" in 8.584s (8.584s including waiting) kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:10 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:10 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:17 +0000 UTC Warning Pod elasticsearch-0.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Get "http://10.129.2.75:9200/": dial tcp 10.129.2.75:9200: connect: connection refused kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:27 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-zookeeper NoPods No matching pods found controllermanager logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:27 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:28 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:28 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-known-ray/data-auto-provisioned-zookeeper-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-754f7b757-ks45n_0febbce6-6aa5-41e9-b1e6-076031582c4f logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:32 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ProvisioningSucceeded Successfully provisioned volume pvc-838194bc-d239-47f9-8cf6-bef9bc84f5e0 ebs.csi.aws.com_aws-ebs-csi-driver-controller-754f7b757-ks45n_0febbce6-6aa5-41e9-b1e6-076031582c4f logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:33 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 Binding Scheduled Successfully assigned kuttl-test-known-ray/auto-provisioned-zookeeper-0 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:34 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-838194bc-d239-47f9-8cf6-bef9bc84f5e0" attachdetach-controller logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:40 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 AddedInterface Add eth0 [10.128.2.81/23] from ovn-kubernetes logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:40 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:55b18000b2e774adccd0d5393b58fc3b8358ab174eaaa3ada9d520b5cde30669" kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:49 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:55b18000b2e774adccd0d5393b58fc3b8358ab174eaaa3ada9d520b5cde30669" in 9.166s (9.166s including waiting) kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:49 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Created Created container zookeeper kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:14:49 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Started Started container zookeeper kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:11 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:12 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-kafka NoPods No matching pods found controllermanager logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:12 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:12 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-known-ray/data-0-auto-provisioned-kafka-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-754f7b757-ks45n_0febbce6-6aa5-41e9-b1e6-076031582c4f logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:16 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ProvisioningSucceeded Successfully provisioned volume pvc-81dcd982-4036-4a73-9573-79d0fb85458e ebs.csi.aws.com_aws-ebs-csi-driver-controller-754f7b757-ks45n_0febbce6-6aa5-41e9-b1e6-076031582c4f logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:17 +0000 UTC Normal Pod auto-provisioned-kafka-0 Binding Scheduled Successfully assigned kuttl-test-known-ray/auto-provisioned-kafka-0 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:19 +0000 UTC Normal Pod auto-provisioned-kafka-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-81dcd982-4036-4a73-9573-79d0fb85458e" attachdetach-controller logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:20 +0000 UTC Normal Pod auto-provisioned-kafka-0 AddedInterface Add eth0 [10.128.2.82/23] from ovn-kubernetes logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:20 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulled Container image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:55b18000b2e774adccd0d5393b58fc3b8358ab174eaaa3ada9d520b5cde30669" already present on machine kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:20 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Created Created container kafka kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:20 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Started Started container kafka kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:42 +0000 UTC Normal Pod auto-provisioned-entity-operator-5dc89dbc55-2mrfn Binding Scheduled Successfully assigned kuttl-test-known-ray/auto-provisioned-entity-operator-5dc89dbc55-2mrfn to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:42 +0000 UTC Normal ReplicaSet.apps auto-provisioned-entity-operator-5dc89dbc55 SuccessfulCreate Created pod: auto-provisioned-entity-operator-5dc89dbc55-2mrfn replicaset-controller logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:42 +0000 UTC Normal Deployment.apps auto-provisioned-entity-operator ScalingReplicaSet Scaled up replica set auto-provisioned-entity-operator-5dc89dbc55 to 1 deployment-controller logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:43 +0000 UTC Normal Pod auto-provisioned-entity-operator-5dc89dbc55-2mrfn AddedInterface Add eth0 [10.129.2.77/23] from ovn-kubernetes logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:43 +0000 UTC Normal Pod auto-provisioned-entity-operator-5dc89dbc55-2mrfn.spec.containers{topic-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:f1be1aa2f18276f9169893eb55e3733cd52fa38f2101a9b3925f79774841689f" already present on machine kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:43 +0000 UTC Normal Pod auto-provisioned-entity-operator-5dc89dbc55-2mrfn.spec.containers{topic-operator} Created Created container topic-operator kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:43 +0000 UTC Normal Pod auto-provisioned-entity-operator-5dc89dbc55-2mrfn.spec.containers{topic-operator} Started Started container topic-operator kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:43 +0000 UTC Normal Pod auto-provisioned-entity-operator-5dc89dbc55-2mrfn.spec.containers{user-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:f1be1aa2f18276f9169893eb55e3733cd52fa38f2101a9b3925f79774841689f" already present on machine kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:43 +0000 UTC Normal Pod auto-provisioned-entity-operator-5dc89dbc55-2mrfn.spec.containers{user-operator} Created Created container user-operator kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:43 +0000 UTC Normal Pod auto-provisioned-entity-operator-5dc89dbc55-2mrfn.spec.containers{user-operator} Started Started container user-operator kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:43 +0000 UTC Normal Pod auto-provisioned-entity-operator-5dc89dbc55-2mrfn.spec.containers{tls-sidecar} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:55b18000b2e774adccd0d5393b58fc3b8358ab174eaaa3ada9d520b5cde30669" kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:52 +0000 UTC Normal Pod auto-provisioned-entity-operator-5dc89dbc55-2mrfn.spec.containers{tls-sidecar} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:55b18000b2e774adccd0d5393b58fc3b8358ab174eaaa3ada9d520b5cde30669" in 8.982s (8.982s including waiting) kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:52 +0000 UTC Normal Pod auto-provisioned-entity-operator-5dc89dbc55-2mrfn.spec.containers{tls-sidecar} Created Created container tls-sidecar kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:15:52 +0000 UTC Normal Pod auto-provisioned-entity-operator-5dc89dbc55-2mrfn.spec.containers{tls-sidecar} Started Started container tls-sidecar kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:15 +0000 UTC Normal Pod auto-provisioned-collector-f7ff84b7b-4wg8z Binding Scheduled Successfully assigned kuttl-test-known-ray/auto-provisioned-collector-f7ff84b7b-4wg8z to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:15 +0000 UTC Warning Pod auto-provisioned-collector-f7ff84b7b-4wg8z FailedMount MountVolume.SetUp failed for volume "auto-provisioned-collector-tls-config-volume" : secret "auto-provisioned-collector-headless-tls" not found kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:15 +0000 UTC Normal ReplicaSet.apps auto-provisioned-collector-f7ff84b7b SuccessfulCreate Created pod: auto-provisioned-collector-f7ff84b7b-4wg8z replicaset-controller logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:15 +0000 UTC Normal Deployment.apps auto-provisioned-collector ScalingReplicaSet Scaled up replica set auto-provisioned-collector-f7ff84b7b to 1 deployment-controller logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:15 +0000 UTC Normal Pod auto-provisioned-ingester-5955774654-grtvm Binding Scheduled Successfully assigned kuttl-test-known-ray/auto-provisioned-ingester-5955774654-grtvm to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:15 +0000 UTC Normal ReplicaSet.apps auto-provisioned-ingester-5955774654 SuccessfulCreate Created pod: auto-provisioned-ingester-5955774654-grtvm replicaset-controller logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:15 +0000 UTC Normal Deployment.apps auto-provisioned-ingester ScalingReplicaSet Scaled up replica set auto-provisioned-ingester-5955774654 to 1 deployment-controller logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:15 +0000 UTC Normal Pod auto-provisioned-query-74fcbb89f8-2cmb8 Binding Scheduled Successfully assigned kuttl-test-known-ray/auto-provisioned-query-74fcbb89f8-2cmb8 to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:15 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-74fcbb89f8 SuccessfulCreate Created pod: auto-provisioned-query-74fcbb89f8-2cmb8 replicaset-controller logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:15 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled up replica set auto-provisioned-query-74fcbb89f8 to 1 deployment-controller logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:16 +0000 UTC Normal Pod auto-provisioned-collector-f7ff84b7b-4wg8z AddedInterface Add eth0 [10.131.0.63/23] from ovn-kubernetes logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:16 +0000 UTC Normal Pod auto-provisioned-collector-f7ff84b7b-4wg8z.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:16 +0000 UTC Normal Pod auto-provisioned-collector-f7ff84b7b-4wg8z.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:16 +0000 UTC Normal Pod auto-provisioned-collector-f7ff84b7b-4wg8z.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:16 +0000 UTC Normal Pod auto-provisioned-ingester-5955774654-grtvm AddedInterface Add eth0 [10.128.2.83/23] from ovn-kubernetes logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:16 +0000 UTC Normal Pod auto-provisioned-ingester-5955774654-grtvm.spec.containers{jaeger-ingester} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:849018528225b7370cc4740fc9f94bef7ffd4195328a916a6013d88f885eebe2" kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:16 +0000 UTC Normal Pod auto-provisioned-query-74fcbb89f8-2cmb8 AddedInterface Add eth0 [10.131.0.64/23] from ovn-kubernetes logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:16 +0000 UTC Normal Pod auto-provisioned-query-74fcbb89f8-2cmb8.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:16 +0000 UTC Normal Pod auto-provisioned-query-74fcbb89f8-2cmb8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:16 +0000 UTC Normal Pod auto-provisioned-query-74fcbb89f8-2cmb8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:16 +0000 UTC Normal Pod auto-provisioned-query-74fcbb89f8-2cmb8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:16 +0000 UTC Normal Pod auto-provisioned-query-74fcbb89f8-2cmb8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:16 +0000 UTC Normal Pod auto-provisioned-query-74fcbb89f8-2cmb8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:16 +0000 UTC Normal Pod auto-provisioned-query-74fcbb89f8-2cmb8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:16 +0000 UTC Normal Pod auto-provisioned-query-74fcbb89f8-2cmb8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:16 +0000 UTC Normal Pod auto-provisioned-query-74fcbb89f8-2cmb8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:22 +0000 UTC Normal Pod auto-provisioned-ingester-5955774654-grtvm.spec.containers{jaeger-ingester} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:849018528225b7370cc4740fc9f94bef7ffd4195328a916a6013d88f885eebe2" in 6.032s (6.032s including waiting) kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:22 +0000 UTC Normal Pod auto-provisioned-ingester-5955774654-grtvm.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | 2023-12-25 08:16:22 +0000 UTC Normal Pod auto-provisioned-ingester-5955774654-grtvm.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 08:16:25 | streaming-with-autoprovisioning-autoscale | Deleting namespace: kuttl-test-known-ray === CONT kuttl/harness/streaming-with-tls logger.go:42: 08:17:01 | streaming-with-tls | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:17:01 | streaming-with-tls | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:17:01 | streaming-with-tls | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:17:01 | streaming-with-tls | Creating namespace: kuttl-test-outgoing-mink logger.go:42: 08:17:01 | streaming-with-tls/0-install | starting test step 0-install logger.go:42: 08:17:01 | streaming-with-tls/0-install | running command: [sh -c cd /tmp/jaeger-tests && make undeploy-kafka KAFKA_NAMESPACE=$NAMESPACE KAFKA_OLM=true] logger.go:42: 08:17:01 | streaming-with-tls/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 08:17:01 | streaming-with-tls/0-install | >>>> Skiping kafka-operator undeploy logger.go:42: 08:17:01 | streaming-with-tls/0-install | kubectl delete --namespace kuttl-test-outgoing-mink -f tests/_build/kafka-example.yaml 2>&1 || true logger.go:42: 08:17:01 | streaming-with-tls/0-install | error: the path "tests/_build/kafka-example.yaml" does not exist logger.go:42: 08:17:01 | streaming-with-tls/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 08:17:01 | streaming-with-tls/0-install | running command: [sh -c cd /tmp/jaeger-tests && make kafka KAFKA_NAMESPACE=$NAMESPACE KAFKA_OLM=true KAFKA_VERSION=3.6.0] logger.go:42: 08:17:01 | streaming-with-tls/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 08:17:01 | streaming-with-tls/0-install | >>>> Creating namespace kuttl-test-outgoing-mink logger.go:42: 08:17:01 | streaming-with-tls/0-install | kubectl create namespace kuttl-test-outgoing-mink 2>&1 | grep -v "already exists" || true logger.go:42: 08:17:01 | streaming-with-tls/0-install | >>>> Skipping kafka-operator deployment, assuming it has been installed via OperatorHub logger.go:42: 08:17:01 | streaming-with-tls/0-install | >>>> Creating namespace kuttl-test-outgoing-mink logger.go:42: 08:17:01 | streaming-with-tls/0-install | mkdir -p tests/_build/ logger.go:42: 08:17:01 | streaming-with-tls/0-install | kubectl create namespace kuttl-test-outgoing-mink 2>&1 | grep -v "already exists" || true logger.go:42: 08:17:02 | streaming-with-tls/0-install | curl --fail --location "https://raw.githubusercontent.com/strimzi/strimzi-kafka-operator/3.6.0/examples/kafka/kafka-persistent-single.yaml" --output tests/_build/kafka-example.yaml --create-dirs logger.go:42: 08:17:02 | streaming-with-tls/0-install | % Total % Received % Xferd Average Speed Time Time Time Current logger.go:42: 08:17:02 | streaming-with-tls/0-install | Dload Upload Total Spent Left Speed logger.go:42: 08:17:02 | streaming-with-tls/0-install | 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 14 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 logger.go:42: 08:17:02 | streaming-with-tls/0-install | curl: (22) The requested URL returned error: 404 logger.go:42: 08:17:02 | streaming-with-tls/0-install | make[2]: *** [Makefile:252: kafka] Error 22 logger.go:42: 08:17:02 | streaming-with-tls/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' case.go:364: failed in step 0-install case.go:366: exit status 2 logger.go:42: 08:17:02 | streaming-with-tls | streaming-with-tls events from ns kuttl-test-outgoing-mink: logger.go:42: 08:17:02 | streaming-with-tls | Deleting namespace: kuttl-test-outgoing-mink === CONT kuttl/harness/streaming-simple logger.go:42: 08:17:08 | streaming-simple | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:17:08 | streaming-simple | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:17:08 | streaming-simple | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:17:08 | streaming-simple | Creating namespace: kuttl-test-live-duck logger.go:42: 08:17:08 | streaming-simple/0-install | starting test step 0-install logger.go:42: 08:17:08 | streaming-simple/0-install | running command: [sh -c cd /tmp/jaeger-tests && make undeploy-kafka KAFKA_NAMESPACE=$NAMESPACE KAFKA_OLM=true] logger.go:42: 08:17:08 | streaming-simple/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 08:17:08 | streaming-simple/0-install | >>>> Skiping kafka-operator undeploy logger.go:42: 08:17:08 | streaming-simple/0-install | kubectl delete --namespace kuttl-test-live-duck -f tests/_build/kafka-example.yaml 2>&1 || true logger.go:42: 08:17:08 | streaming-simple/0-install | error: the path "tests/_build/kafka-example.yaml" does not exist logger.go:42: 08:17:08 | streaming-simple/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 08:17:08 | streaming-simple/0-install | running command: [sh -c cd /tmp/jaeger-tests && make kafka KAFKA_NAMESPACE=$NAMESPACE KAFKA_OLM=true KAFKA_VERSION=3.6.0] logger.go:42: 08:17:08 | streaming-simple/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 08:17:08 | streaming-simple/0-install | >>>> Creating namespace kuttl-test-live-duck logger.go:42: 08:17:08 | streaming-simple/0-install | kubectl create namespace kuttl-test-live-duck 2>&1 | grep -v "already exists" || true logger.go:42: 08:17:08 | streaming-simple/0-install | >>>> Skipping kafka-operator deployment, assuming it has been installed via OperatorHub logger.go:42: 08:17:08 | streaming-simple/0-install | >>>> Creating namespace kuttl-test-live-duck logger.go:42: 08:17:08 | streaming-simple/0-install | mkdir -p tests/_build/ logger.go:42: 08:17:08 | streaming-simple/0-install | kubectl create namespace kuttl-test-live-duck 2>&1 | grep -v "already exists" || true logger.go:42: 08:17:09 | streaming-simple/0-install | curl --fail --location "https://raw.githubusercontent.com/strimzi/strimzi-kafka-operator/3.6.0/examples/kafka/kafka-persistent-single.yaml" --output tests/_build/kafka-example.yaml --create-dirs logger.go:42: 08:17:09 | streaming-simple/0-install | % Total % Received % Xferd Average Speed Time Time Time Current logger.go:42: 08:17:09 | streaming-simple/0-install | Dload Upload Total Spent Left Speed logger.go:42: 08:17:09 | streaming-simple/0-install | 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 14 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 logger.go:42: 08:17:09 | streaming-simple/0-install | curl: (22) The requested URL returned error: 404 logger.go:42: 08:17:09 | streaming-simple/0-install | make[2]: *** [Makefile:252: kafka] Error 22 logger.go:42: 08:17:09 | streaming-simple/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' case.go:364: failed in step 0-install case.go:366: exit status 2 logger.go:42: 08:17:09 | streaming-simple | streaming-simple events from ns kuttl-test-live-duck: logger.go:42: 08:17:09 | streaming-simple | Deleting namespace: kuttl-test-live-duck === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (206.80s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.23s) --- PASS: kuttl/harness/streaming-with-autoprovisioning-autoscale (186.68s) --- FAIL: kuttl/harness/streaming-with-tls (7.06s) --- FAIL: kuttl/harness/streaming-simple (6.67s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name streaming --report --output /logs/artifacts/streaming.xml ./artifacts/kuttl-report.xml time="2023-12-25T08:17:16Z" level=debug msg="Setting a new name for the test suites" time="2023-12-25T08:17:16Z" level=debug msg="Removing 'artifacts' TestCase" time="2023-12-25T08:17:16Z" level=debug msg="normalizing test case names" time="2023-12-25T08:17:16Z" level=debug msg="streaming/artifacts -> streaming_artifacts" time="2023-12-25T08:17:16Z" level=debug msg="streaming/streaming-with-autoprovisioning-autoscale -> streaming_streaming_with_autoprovisioning_autoscale" time="2023-12-25T08:17:16Z" level=debug msg="streaming/streaming-with-tls -> streaming_streaming_with_tls" time="2023-12-25T08:17:16Z" level=debug msg="streaming/streaming-simple -> streaming_streaming_simple" +-----------------------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------------------+--------+ | streaming_artifacts | passed | | streaming_streaming_with_autoprovisioning_autoscale | passed | | streaming_streaming_with_tls | failed | | streaming_streaming_simple | failed | +-----------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh ui false true + '[' 3 -ne 3 ']' + test_suite_name=ui + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/ui.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-ui make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true ./tests/e2e/ui/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-12-22-165425 True False 51m Cluster version is 4.15.0-0.nightly-2023-12-22-165425' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-12-22-165425 True False 51m Cluster version is 4.15.0-0.nightly-2023-12-22-165425' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/ui/render.sh ++ export SUITE_DIR=./tests/e2e/ui ++ SUITE_DIR=./tests/e2e/ui ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/ui ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test allinone + '[' 1 -ne 1 ']' + test_name=allinone + echo =========================================================================== =========================================================================== + info 'Rendering files for test allinone' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test allinone\e[0m' Rendering files for test allinone + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build + '[' _build '!=' _build ']' + mkdir -p allinone + cd allinone + export GET_URL_COMMAND + export URL + export JAEGER_NAME=all-in-one-ui + JAEGER_NAME=all-in-one-ui + '[' true = true ']' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./01-curl.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./04-test-ui-config.yaml + start_test production + '[' 1 -ne 1 ']' + test_name=production + echo =========================================================================== =========================================================================== + info 'Rendering files for test production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test production\e[0m' Rendering files for test production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build/allinone + '[' allinone '!=' _build ']' + cd .. + mkdir -p production + cd production + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + [[ true = true ]] + [[ true = true ]] + render_install_jaeger production-ui production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + '[' true = true ']' + INSECURE=true + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-forbbiden-access.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-curl.yaml + INSECURE=true + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./05-check-disabled-security.yaml + ASSERT_PRESENT=false + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./06-check-NO-gaID.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./08-check-gaID.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running ui E2E tests' Running ui E2E tests + cd tests/e2e/ui/_build + set +e + KUBECONFIG=/tmp/kubeconfig-1441553849 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 3 tests === RUN kuttl/harness === RUN kuttl/harness/allinone === PAUSE kuttl/harness/allinone === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/production === PAUSE kuttl/harness/production === CONT kuttl/harness/allinone logger.go:42: 08:17:23 | allinone | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:17:23 | allinone | Creating namespace: kuttl-test-settled-molly logger.go:42: 08:17:23 | allinone/0-install | starting test step 0-install logger.go:42: 08:17:23 | allinone/0-install | Jaeger:kuttl-test-settled-molly/all-in-one-ui created logger.go:42: 08:17:27 | allinone/0-install | test step completed 0-install logger.go:42: 08:17:27 | allinone/1-curl | starting test step 1-curl logger.go:42: 08:17:27 | allinone/1-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 08:17:27 | allinone/1-curl | Checking the Ingress host value was populated logger.go:42: 08:17:27 | allinone/1-curl | Try number 0 logger.go:42: 08:17:27 | allinone/1-curl | Hostname is all-in-one-ui-kuttl-test-settled-molly.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com logger.go:42: 08:17:27 | allinone/1-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE all-in-one-ui] logger.go:42: 08:17:27 | allinone/1-curl | Checking an expected HTTP response logger.go:42: 08:17:27 | allinone/1-curl | Running in OpenShift logger.go:42: 08:17:27 | allinone/1-curl | User not provided. Getting the token... logger.go:42: 08:17:29 | allinone/1-curl | Warning: resource jaegers/all-in-one-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:17:35 | allinone/1-curl | Try number 1/30 the https://all-in-one-ui-kuttl-test-settled-molly.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:17:35 | allinone/1-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 08:17:35 | allinone/1-curl | Try number 2/30 the https://all-in-one-ui-kuttl-test-settled-molly.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:17:35 | allinone/1-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 08:17:45 | allinone/1-curl | Try number 3/30 the https://all-in-one-ui-kuttl-test-settled-molly.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:17:45 | allinone/1-curl | curl response asserted properly logger.go:42: 08:17:45 | allinone/1-curl | test step completed 1-curl logger.go:42: 08:17:45 | allinone/2-delete | starting test step 2-delete logger.go:42: 08:17:45 | allinone/2-delete | Jaeger:kuttl-test-settled-molly/all-in-one-ui created logger.go:42: 08:17:45 | allinone/2-delete | test step completed 2-delete logger.go:42: 08:17:45 | allinone/3-install | starting test step 3-install logger.go:42: 08:17:45 | allinone/3-install | Jaeger:kuttl-test-settled-molly/all-in-one-ui updated logger.go:42: 08:17:45 | allinone/3-install | test step completed 3-install logger.go:42: 08:17:45 | allinone/4-test-ui-config | starting test step 4-test-ui-config logger.go:42: 08:17:45 | allinone/4-test-ui-config | running command: [./ensure-ingress-host.sh] logger.go:42: 08:17:45 | allinone/4-test-ui-config | Checking the Ingress host value was populated logger.go:42: 08:17:45 | allinone/4-test-ui-config | Try number 0 logger.go:42: 08:17:45 | allinone/4-test-ui-config | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 08:17:45 | allinone/4-test-ui-config | template was: logger.go:42: 08:17:45 | allinone/4-test-ui-config | {.items[0].status.ingress[0].host} logger.go:42: 08:17:45 | allinone/4-test-ui-config | object given to jsonpath engine was: logger.go:42: 08:17:45 | allinone/4-test-ui-config | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 08:17:45 | allinone/4-test-ui-config | logger.go:42: 08:17:45 | allinone/4-test-ui-config | logger.go:42: 08:17:55 | allinone/4-test-ui-config | Try number 1 logger.go:42: 08:17:56 | allinone/4-test-ui-config | Hostname is all-in-one-ui-kuttl-test-settled-molly.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com logger.go:42: 08:17:56 | allinone/4-test-ui-config | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 08:17:56 | allinone/4-test-ui-config | time="2023-12-25T08:17:56Z" level=info msg="Querying https://all-in-one-ui-kuttl-test-settled-molly.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search..." logger.go:42: 08:17:56 | allinone/4-test-ui-config | time="2023-12-25T08:17:56Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 08:17:56 | allinone/4-test-ui-config | time="2023-12-25T08:17:56Z" level=info msg="Polling to https://all-in-one-ui-kuttl-test-settled-molly.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search" logger.go:42: 08:17:56 | allinone/4-test-ui-config | time="2023-12-25T08:17:56Z" level=info msg="Doing request number 0" logger.go:42: 08:17:56 | allinone/4-test-ui-config | time="2023-12-25T08:17:56Z" level=warning msg="Status code: 503" logger.go:42: 08:18:04 | allinone/4-test-ui-config | time="2023-12-25T08:18:04Z" level=info msg="Doing request number 1" logger.go:42: 08:18:04 | allinone/4-test-ui-config | time="2023-12-25T08:18:04Z" level=info msg="Content found and asserted!" logger.go:42: 08:18:04 | allinone/4-test-ui-config | time="2023-12-25T08:18:04Z" level=info msg="Success!" logger.go:42: 08:18:04 | allinone/4-test-ui-config | test step completed 4-test-ui-config logger.go:42: 08:18:04 | allinone | allinone events from ns kuttl-test-settled-molly: logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:26 +0000 UTC Normal Pod all-in-one-ui-db6ff759f-2v2g2 Binding Scheduled Successfully assigned kuttl-test-settled-molly/all-in-one-ui-db6ff759f-2v2g2 to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:26 +0000 UTC Warning Pod all-in-one-ui-db6ff759f-2v2g2 FailedMount MountVolume.SetUp failed for volume "all-in-one-ui-ui-oauth-proxy-tls" : secret "all-in-one-ui-ui-oauth-proxy-tls" not found kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:26 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-db6ff759f SuccessfulCreate Created pod: all-in-one-ui-db6ff759f-2v2g2 replicaset-controller logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:26 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-db6ff759f to 1 deployment-controller logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:27 +0000 UTC Normal Pod all-in-one-ui-db6ff759f-2v2g2 AddedInterface Add eth0 [10.129.2.78/23] from ovn-kubernetes logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:27 +0000 UTC Normal Pod all-in-one-ui-db6ff759f-2v2g2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:27 +0000 UTC Normal Pod all-in-one-ui-db6ff759f-2v2g2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:27 +0000 UTC Normal Pod all-in-one-ui-db6ff759f-2v2g2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:27 +0000 UTC Normal Pod all-in-one-ui-db6ff759f-2v2g2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:27 +0000 UTC Normal Pod all-in-one-ui-db6ff759f-2v2g2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:27 +0000 UTC Normal Pod all-in-one-ui-db6ff759f-2v2g2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:31 +0000 UTC Normal Pod all-in-one-ui-db6ff759f-2v2g2.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:31 +0000 UTC Normal Pod all-in-one-ui-db6ff759f-2v2g2.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:31 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-db6ff759f SuccessfulDelete Deleted pod: all-in-one-ui-db6ff759f-2v2g2 replicaset-controller logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:31 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-db6ff759f to 0 from 1 deployment-controller logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:32 +0000 UTC Normal Pod all-in-one-ui-f886bdd8f-h7v5z Binding Scheduled Successfully assigned kuttl-test-settled-molly/all-in-one-ui-f886bdd8f-h7v5z to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:32 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-f886bdd8f SuccessfulCreate Created pod: all-in-one-ui-f886bdd8f-h7v5z replicaset-controller logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:32 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-f886bdd8f to 1 deployment-controller logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:33 +0000 UTC Normal Pod all-in-one-ui-f886bdd8f-h7v5z AddedInterface Add eth0 [10.129.2.79/23] from ovn-kubernetes logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:33 +0000 UTC Normal Pod all-in-one-ui-f886bdd8f-h7v5z.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:33 +0000 UTC Normal Pod all-in-one-ui-f886bdd8f-h7v5z.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:33 +0000 UTC Normal Pod all-in-one-ui-f886bdd8f-h7v5z.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:33 +0000 UTC Normal Pod all-in-one-ui-f886bdd8f-h7v5z.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:33 +0000 UTC Normal Pod all-in-one-ui-f886bdd8f-h7v5z.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:33 +0000 UTC Normal Pod all-in-one-ui-f886bdd8f-h7v5z.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:45 +0000 UTC Normal Pod all-in-one-ui-f886bdd8f-h7v5z.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:45 +0000 UTC Normal Pod all-in-one-ui-f886bdd8f-h7v5z.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:51 +0000 UTC Normal Pod all-in-one-ui-fd857fd75-hr42s Binding Scheduled Successfully assigned kuttl-test-settled-molly/all-in-one-ui-fd857fd75-hr42s to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:51 +0000 UTC Normal Pod all-in-one-ui-fd857fd75-hr42s AddedInterface Add eth0 [10.129.2.80/23] from ovn-kubernetes logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:51 +0000 UTC Normal Pod all-in-one-ui-fd857fd75-hr42s.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:51 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-fd857fd75 SuccessfulCreate Created pod: all-in-one-ui-fd857fd75-hr42s replicaset-controller logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:51 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-fd857fd75 to 1 deployment-controller logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:52 +0000 UTC Normal Pod all-in-one-ui-fd857fd75-hr42s.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:18:04 | allinone | 2023-12-25 08:17:52 +0000 UTC Normal Pod all-in-one-ui-fd857fd75-hr42s.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:18:04 | allinone | Deleting namespace: kuttl-test-settled-molly === CONT kuttl/harness/production logger.go:42: 08:18:11 | production | Ignoring add-tracking-id.yaml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:18:11 | production | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:18:11 | production | Creating namespace: kuttl-test-blessed-javelin logger.go:42: 08:18:11 | production/1-install | starting test step 1-install logger.go:42: 08:18:11 | production/1-install | Jaeger:kuttl-test-blessed-javelin/production-ui created logger.go:42: 08:18:46 | production/1-install | test step completed 1-install logger.go:42: 08:18:46 | production/2-check-forbbiden-access | starting test step 2-check-forbbiden-access logger.go:42: 08:18:46 | production/2-check-forbbiden-access | running command: [./ensure-ingress-host.sh] logger.go:42: 08:18:46 | production/2-check-forbbiden-access | Checking the Ingress host value was populated logger.go:42: 08:18:46 | production/2-check-forbbiden-access | Try number 0 logger.go:42: 08:18:46 | production/2-check-forbbiden-access | Hostname is production-ui-kuttl-test-blessed-javelin.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com logger.go:42: 08:18:46 | production/2-check-forbbiden-access | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE production-ui] logger.go:42: 08:18:46 | production/2-check-forbbiden-access | Checking an expected HTTP response logger.go:42: 08:18:46 | production/2-check-forbbiden-access | Running in OpenShift logger.go:42: 08:18:46 | production/2-check-forbbiden-access | Not using any secret logger.go:42: 08:18:46 | production/2-check-forbbiden-access | Try number 1/30 the https://production-ui-kuttl-test-blessed-javelin.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:18:46 | production/2-check-forbbiden-access | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 08:18:46 | production/2-check-forbbiden-access | Try number 2/30 the https://production-ui-kuttl-test-blessed-javelin.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:18:46 | production/2-check-forbbiden-access | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 08:18:56 | production/2-check-forbbiden-access | Try number 3/30 the https://production-ui-kuttl-test-blessed-javelin.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:18:57 | production/2-check-forbbiden-access | curl response asserted properly logger.go:42: 08:18:57 | production/2-check-forbbiden-access | test step completed 2-check-forbbiden-access logger.go:42: 08:18:57 | production/3-curl | starting test step 3-curl logger.go:42: 08:18:57 | production/3-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 08:18:57 | production/3-curl | Checking the Ingress host value was populated logger.go:42: 08:18:57 | production/3-curl | Try number 0 logger.go:42: 08:18:57 | production/3-curl | Hostname is production-ui-kuttl-test-blessed-javelin.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com logger.go:42: 08:18:57 | production/3-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 08:18:57 | production/3-curl | Checking an expected HTTP response logger.go:42: 08:18:57 | production/3-curl | Running in OpenShift logger.go:42: 08:18:57 | production/3-curl | User not provided. Getting the token... logger.go:42: 08:18:58 | production/3-curl | Warning: resource jaegers/production-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:19:05 | production/3-curl | Try number 1/30 the https://production-ui-kuttl-test-blessed-javelin.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:19:05 | production/3-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 08:19:05 | production/3-curl | Try number 2/30 the https://production-ui-kuttl-test-blessed-javelin.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:19:05 | production/3-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 08:19:15 | production/3-curl | Try number 3/30 the https://production-ui-kuttl-test-blessed-javelin.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:19:15 | production/3-curl | curl response asserted properly logger.go:42: 08:19:15 | production/3-curl | test step completed 3-curl logger.go:42: 08:19:15 | production/4-install | starting test step 4-install logger.go:42: 08:19:15 | production/4-install | Jaeger:kuttl-test-blessed-javelin/production-ui updated logger.go:42: 08:19:15 | production/4-install | test step completed 4-install logger.go:42: 08:19:15 | production/5-check-disabled-security | starting test step 5-check-disabled-security logger.go:42: 08:19:15 | production/5-check-disabled-security | running command: [./ensure-ingress-host.sh] logger.go:42: 08:19:15 | production/5-check-disabled-security | Checking the Ingress host value was populated logger.go:42: 08:19:15 | production/5-check-disabled-security | Try number 0 logger.go:42: 08:19:15 | production/5-check-disabled-security | Hostname is production-ui-kuttl-test-blessed-javelin.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com logger.go:42: 08:19:15 | production/5-check-disabled-security | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 08:19:15 | production/5-check-disabled-security | Checking an expected HTTP response logger.go:42: 08:19:15 | production/5-check-disabled-security | Running in OpenShift logger.go:42: 08:19:15 | production/5-check-disabled-security | Not using any secret logger.go:42: 08:19:15 | production/5-check-disabled-security | Try number 1/30 the https://production-ui-kuttl-test-blessed-javelin.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:19:15 | production/5-check-disabled-security | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 08:19:15 | production/5-check-disabled-security | Try number 2/30 the https://production-ui-kuttl-test-blessed-javelin.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:19:15 | production/5-check-disabled-security | HTTP response is 403. 200 expected. Waiting 10 s logger.go:42: 08:19:25 | production/5-check-disabled-security | Try number 3/30 the https://production-ui-kuttl-test-blessed-javelin.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 08:19:25 | production/5-check-disabled-security | curl response asserted properly logger.go:42: 08:19:25 | production/5-check-disabled-security | test step completed 5-check-disabled-security logger.go:42: 08:19:25 | production/6-check-NO-gaID | starting test step 6-check-NO-gaID logger.go:42: 08:19:25 | production/6-check-NO-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 08:19:25 | production/6-check-NO-gaID | Checking the Ingress host value was populated logger.go:42: 08:19:25 | production/6-check-NO-gaID | Try number 0 logger.go:42: 08:19:25 | production/6-check-NO-gaID | Hostname is production-ui-kuttl-test-blessed-javelin.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com logger.go:42: 08:19:25 | production/6-check-NO-gaID | running command: [sh -c ASSERT_PRESENT=false EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 08:19:26 | production/6-check-NO-gaID | time="2023-12-25T08:19:26Z" level=info msg="Querying https://production-ui-kuttl-test-blessed-javelin.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search..." logger.go:42: 08:19:26 | production/6-check-NO-gaID | time="2023-12-25T08:19:26Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 08:19:26 | production/6-check-NO-gaID | time="2023-12-25T08:19:26Z" level=info msg="Polling to https://production-ui-kuttl-test-blessed-javelin.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search" logger.go:42: 08:19:26 | production/6-check-NO-gaID | time="2023-12-25T08:19:26Z" level=info msg="Doing request number 0" logger.go:42: 08:19:26 | production/6-check-NO-gaID | time="2023-12-25T08:19:26Z" level=info msg="Content not found and asserted it was not found!" logger.go:42: 08:19:26 | production/6-check-NO-gaID | time="2023-12-25T08:19:26Z" level=info msg="Success!" logger.go:42: 08:19:26 | production/6-check-NO-gaID | test step completed 6-check-NO-gaID logger.go:42: 08:19:26 | production/7-add-tracking-id | starting test step 7-add-tracking-id logger.go:42: 08:19:26 | production/7-add-tracking-id | running command: [sh -c kubectl apply -f add-tracking-id.yaml -n $NAMESPACE] logger.go:42: 08:19:26 | production/7-add-tracking-id | jaeger.jaegertracing.io/production-ui configured logger.go:42: 08:19:26 | production/7-add-tracking-id | test step completed 7-add-tracking-id logger.go:42: 08:19:26 | production/8-check-gaID | starting test step 8-check-gaID logger.go:42: 08:19:26 | production/8-check-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 08:19:26 | production/8-check-gaID | Checking the Ingress host value was populated logger.go:42: 08:19:26 | production/8-check-gaID | Try number 0 logger.go:42: 08:19:26 | production/8-check-gaID | Hostname is production-ui-kuttl-test-blessed-javelin.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com logger.go:42: 08:19:26 | production/8-check-gaID | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 08:19:27 | production/8-check-gaID | time="2023-12-25T08:19:27Z" level=info msg="Querying https://production-ui-kuttl-test-blessed-javelin.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search..." logger.go:42: 08:19:27 | production/8-check-gaID | time="2023-12-25T08:19:27Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 08:19:27 | production/8-check-gaID | time="2023-12-25T08:19:27Z" level=info msg="Polling to https://production-ui-kuttl-test-blessed-javelin.apps.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com/search" logger.go:42: 08:19:27 | production/8-check-gaID | time="2023-12-25T08:19:27Z" level=info msg="Doing request number 0" logger.go:42: 08:19:27 | production/8-check-gaID | time="2023-12-25T08:19:27Z" level=warning msg="Found: false . Assert: true" logger.go:42: 08:19:27 | production/8-check-gaID | time="2023-12-25T08:19:27Z" level=warning msg="The condition of the test function was not accomplished" logger.go:42: 08:19:35 | production/8-check-gaID | time="2023-12-25T08:19:35Z" level=info msg="Doing request number 1" logger.go:42: 08:19:35 | production/8-check-gaID | time="2023-12-25T08:19:35Z" level=info msg="Content found and asserted!" logger.go:42: 08:19:35 | production/8-check-gaID | time="2023-12-25T08:19:35Z" level=info msg="Success!" logger.go:42: 08:19:35 | production/8-check-gaID | test step completed 8-check-gaID logger.go:42: 08:19:35 | production | production events from ns kuttl-test-blessed-javelin: logger.go:42: 08:19:35 | production | 2023-12-25 08:18:17 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestblessedjavelinproductionui-1-546d6fd7bf SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestblessedjavelinproductionui-1-54sfgsb replicaset-controller logger.go:42: 08:19:35 | production | 2023-12-25 08:18:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestblessedjavelinproductionui-1-54sfgsb Binding Scheduled Successfully assigned kuttl-test-blessed-javelin/elasticsearch-cdm-kuttltestblessedjavelinproductionui-1-54sfgsb to ip-10-0-38-209.us-east-2.compute.internal default-scheduler logger.go:42: 08:19:35 | production | 2023-12-25 08:18:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestblessedjavelinproductionui-1-54sfgsb AddedInterface Add eth0 [10.129.2.81/23] from ovn-kubernetes logger.go:42: 08:19:35 | production | 2023-12-25 08:18:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestblessedjavelinproductionui-1-54sfgsb.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestblessedjavelinproductionui-1-54sfgsb.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestblessedjavelinproductionui-1-54sfgsb.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestblessedjavelinproductionui-1-54sfgsb.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestblessedjavelinproductionui-1-54sfgsb.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestblessedjavelinproductionui-1-54sfgsb.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:17 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestblessedjavelinproductionui-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestblessedjavelinproductionui-1-546d6fd7bf to 1 deployment-controller logger.go:42: 08:19:35 | production | 2023-12-25 08:18:27 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestblessedjavelinproductionui-1-54sfgsb.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:32 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestblessedjavelinproductionui-1-54sfgsb.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal Pod production-ui-collector-6b4b9497d9-xsch7 Binding Scheduled Successfully assigned kuttl-test-blessed-javelin/production-ui-collector-6b4b9497d9-xsch7 to ip-10-0-98-142.us-east-2.compute.internal default-scheduler logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal Pod production-ui-collector-6b4b9497d9-xsch7 AddedInterface Add eth0 [10.128.2.84/23] from ovn-kubernetes logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal Pod production-ui-collector-6b4b9497d9-xsch7.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal Pod production-ui-collector-6b4b9497d9-xsch7.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal Pod production-ui-collector-6b4b9497d9-xsch7.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal ReplicaSet.apps production-ui-collector-6b4b9497d9 SuccessfulCreate Created pod: production-ui-collector-6b4b9497d9-xsch7 replicaset-controller logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal Deployment.apps production-ui-collector ScalingReplicaSet Scaled up replica set production-ui-collector-6b4b9497d9 to 1 deployment-controller logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal Pod production-ui-query-5c4f8cb96b-5rddf Binding Scheduled Successfully assigned kuttl-test-blessed-javelin/production-ui-query-5c4f8cb96b-5rddf to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal Pod production-ui-query-5c4f8cb96b-5rddf AddedInterface Add eth0 [10.131.0.65/23] from ovn-kubernetes logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal Pod production-ui-query-5c4f8cb96b-5rddf.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal Pod production-ui-query-5c4f8cb96b-5rddf.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal Pod production-ui-query-5c4f8cb96b-5rddf.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal Pod production-ui-query-5c4f8cb96b-5rddf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal Pod production-ui-query-5c4f8cb96b-5rddf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal Pod production-ui-query-5c4f8cb96b-5rddf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal Pod production-ui-query-5c4f8cb96b-5rddf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal Pod production-ui-query-5c4f8cb96b-5rddf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal Pod production-ui-query-5c4f8cb96b-5rddf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal ReplicaSet.apps production-ui-query-5c4f8cb96b SuccessfulCreate Created pod: production-ui-query-5c4f8cb96b-5rddf replicaset-controller logger.go:42: 08:19:35 | production | 2023-12-25 08:18:44 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-5c4f8cb96b to 1 deployment-controller logger.go:42: 08:19:35 | production | 2023-12-25 08:18:59 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:19:35 | production | 2023-12-25 08:18:59 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:19:35 | production | 2023-12-25 08:18:59 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:19:35 | production | 2023-12-25 08:19:00 +0000 UTC Normal Pod production-ui-query-5c4f8cb96b-5rddf.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:00 +0000 UTC Normal Pod production-ui-query-5c4f8cb96b-5rddf.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:00 +0000 UTC Normal Pod production-ui-query-5c4f8cb96b-5rddf.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:00 +0000 UTC Normal ReplicaSet.apps production-ui-query-5c4f8cb96b SuccessfulDelete Deleted pod: production-ui-query-5c4f8cb96b-5rddf replicaset-controller logger.go:42: 08:19:35 | production | 2023-12-25 08:19:00 +0000 UTC Normal Pod production-ui-query-644b97b5d8-8rcxb Binding Scheduled Successfully assigned kuttl-test-blessed-javelin/production-ui-query-644b97b5d8-8rcxb to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:19:35 | production | 2023-12-25 08:19:00 +0000 UTC Normal ReplicaSet.apps production-ui-query-644b97b5d8 SuccessfulCreate Created pod: production-ui-query-644b97b5d8-8rcxb replicaset-controller logger.go:42: 08:19:35 | production | 2023-12-25 08:19:00 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-5c4f8cb96b to 0 from 1 deployment-controller logger.go:42: 08:19:35 | production | 2023-12-25 08:19:00 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-644b97b5d8 to 1 deployment-controller logger.go:42: 08:19:35 | production | 2023-12-25 08:19:01 +0000 UTC Normal Pod production-ui-query-644b97b5d8-8rcxb AddedInterface Add eth0 [10.131.0.66/23] from ovn-kubernetes logger.go:42: 08:19:35 | production | 2023-12-25 08:19:01 +0000 UTC Normal Pod production-ui-query-644b97b5d8-8rcxb.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:01 +0000 UTC Normal Pod production-ui-query-644b97b5d8-8rcxb.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:01 +0000 UTC Normal Pod production-ui-query-644b97b5d8-8rcxb.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:01 +0000 UTC Normal Pod production-ui-query-644b97b5d8-8rcxb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:01 +0000 UTC Normal Pod production-ui-query-644b97b5d8-8rcxb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:01 +0000 UTC Normal Pod production-ui-query-644b97b5d8-8rcxb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:01 +0000 UTC Normal Pod production-ui-query-644b97b5d8-8rcxb.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:01 +0000 UTC Normal Pod production-ui-query-644b97b5d8-8rcxb.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:01 +0000 UTC Normal Pod production-ui-query-644b97b5d8-8rcxb.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:16 +0000 UTC Normal Pod production-ui-query-644b97b5d8-8rcxb.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:16 +0000 UTC Normal Pod production-ui-query-644b97b5d8-8rcxb.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:16 +0000 UTC Normal Pod production-ui-query-644b97b5d8-8rcxb.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:16 +0000 UTC Normal ReplicaSet.apps production-ui-query-644b97b5d8 SuccessfulDelete Deleted pod: production-ui-query-644b97b5d8-8rcxb replicaset-controller logger.go:42: 08:19:35 | production | 2023-12-25 08:19:16 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-644b97b5d8 to 0 from 1 deployment-controller logger.go:42: 08:19:35 | production | 2023-12-25 08:19:17 +0000 UTC Normal Pod production-ui-query-5c8595564f-k6rkh Binding Scheduled Successfully assigned kuttl-test-blessed-javelin/production-ui-query-5c8595564f-k6rkh to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:19:35 | production | 2023-12-25 08:19:17 +0000 UTC Normal ReplicaSet.apps production-ui-query-5c8595564f SuccessfulCreate Created pod: production-ui-query-5c8595564f-k6rkh replicaset-controller logger.go:42: 08:19:35 | production | 2023-12-25 08:19:17 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-5c8595564f to 1 deployment-controller logger.go:42: 08:19:35 | production | 2023-12-25 08:19:18 +0000 UTC Normal Pod production-ui-query-5c8595564f-k6rkh AddedInterface Add eth0 [10.131.0.67/23] from ovn-kubernetes logger.go:42: 08:19:35 | production | 2023-12-25 08:19:18 +0000 UTC Normal Pod production-ui-query-5c8595564f-k6rkh.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:18 +0000 UTC Normal Pod production-ui-query-5c8595564f-k6rkh.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:18 +0000 UTC Normal Pod production-ui-query-5c8595564f-k6rkh.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:18 +0000 UTC Normal Pod production-ui-query-5c8595564f-k6rkh.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:18 +0000 UTC Normal Pod production-ui-query-5c8595564f-k6rkh.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:18 +0000 UTC Normal Pod production-ui-query-5c8595564f-k6rkh.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:27 +0000 UTC Normal Pod production-ui-query-5c8595564f-k6rkh.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:27 +0000 UTC Normal Pod production-ui-query-5c8595564f-k6rkh.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:27 +0000 UTC Normal ReplicaSet.apps production-ui-query-5c8595564f SuccessfulDelete Deleted pod: production-ui-query-5c8595564f-k6rkh replicaset-controller logger.go:42: 08:19:35 | production | 2023-12-25 08:19:27 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-5c8595564f to 0 from 1 deployment-controller logger.go:42: 08:19:35 | production | 2023-12-25 08:19:28 +0000 UTC Normal Pod production-ui-query-f84c5b966-jbkvl Binding Scheduled Successfully assigned kuttl-test-blessed-javelin/production-ui-query-f84c5b966-jbkvl to ip-10-0-81-222.us-east-2.compute.internal default-scheduler logger.go:42: 08:19:35 | production | 2023-12-25 08:19:28 +0000 UTC Normal ReplicaSet.apps production-ui-query-f84c5b966 SuccessfulCreate Created pod: production-ui-query-f84c5b966-jbkvl replicaset-controller logger.go:42: 08:19:35 | production | 2023-12-25 08:19:28 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-f84c5b966 to 1 deployment-controller logger.go:42: 08:19:35 | production | 2023-12-25 08:19:29 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 08:19:35 | production | 2023-12-25 08:19:29 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod production-ui-collector-6b4b9497d9-xsch7 horizontal-pod-autoscaler logger.go:42: 08:19:35 | production | 2023-12-25 08:19:29 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 08:19:35 | production | 2023-12-25 08:19:29 +0000 UTC Warning Pod production-ui-query-5c8595564f-k6rkh.spec.containers{jaeger-query} Unhealthy Readiness probe failed: Get "http://10.131.0.67:16687/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:29 +0000 UTC Warning Pod production-ui-query-5c8595564f-k6rkh.spec.containers{jaeger-agent} Unhealthy Readiness probe failed: Get "http://10.131.0.67:14271/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:29 +0000 UTC Normal Pod production-ui-query-f84c5b966-jbkvl AddedInterface Add eth0 [10.131.0.68/23] from ovn-kubernetes logger.go:42: 08:19:35 | production | 2023-12-25 08:19:29 +0000 UTC Normal Pod production-ui-query-f84c5b966-jbkvl.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:29 +0000 UTC Normal Pod production-ui-query-f84c5b966-jbkvl.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:29 +0000 UTC Normal Pod production-ui-query-f84c5b966-jbkvl.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:29 +0000 UTC Normal Pod production-ui-query-f84c5b966-jbkvl.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:29 +0000 UTC Normal Pod production-ui-query-f84c5b966-jbkvl.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:19:35 | production | 2023-12-25 08:19:29 +0000 UTC Normal Pod production-ui-query-f84c5b966-jbkvl.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:19:35 | production | Deleting namespace: kuttl-test-blessed-javelin === CONT kuttl/harness/artifacts logger.go:42: 08:19:42 | artifacts | Creating namespace: kuttl-test-apparent-foxhound logger.go:42: 08:19:42 | artifacts | artifacts events from ns kuttl-test-apparent-foxhound: logger.go:42: 08:19:42 | artifacts | Deleting namespace: kuttl-test-apparent-foxhound === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (145.87s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/allinone (48.24s) --- PASS: kuttl/harness/production (91.17s) --- PASS: kuttl/harness/artifacts (6.30s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name ui --report --output /logs/artifacts/ui.xml ./artifacts/kuttl-report.xml time="2023-12-25T08:19:49Z" level=debug msg="Setting a new name for the test suites" time="2023-12-25T08:19:49Z" level=debug msg="Removing 'artifacts' TestCase" time="2023-12-25T08:19:49Z" level=debug msg="normalizing test case names" time="2023-12-25T08:19:49Z" level=debug msg="ui/allinone -> ui_allinone" time="2023-12-25T08:19:49Z" level=debug msg="ui/production -> ui_production" time="2023-12-25T08:19:49Z" level=debug msg="ui/artifacts -> ui_artifacts" +---------------+--------+ | NAME | RESULT | +---------------+--------+ | ui_allinone | passed | | ui_production | passed | | ui_artifacts | passed | +---------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh upgrade false true + '[' 3 -ne 3 ']' + test_suite_name=upgrade + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/upgrade.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-upgrade make[2]: Entering directory '/tmp/jaeger-tests' make docker JAEGER_VERSION=1.52.1 IMG="quay.io//jaeger-operator:next" make[3]: Entering directory '/tmp/jaeger-tests' [ ! -z "true" ] || docker build --build-arg=GOPROXY= --build-arg=VERSION="1.52.0" --build-arg=JAEGER_VERSION=1.52.1 --build-arg=TARGETARCH= --build-arg VERSION_DATE=2023-12-25T08:19:49Z --build-arg VERSION_PKG="github.com/jaegertracing/jaeger-operator/pkg/version" -t "quay.io//jaeger-operator:next" . make[3]: Leaving directory '/tmp/jaeger-tests' touch build-e2e-upgrade-image SKIP_ES_EXTERNAL=true IMG=quay.io//jaeger-operator:"1.52.0" JAEGER_OPERATOR_VERSION="1.52.0" JAEGER_VERSION="1.52.0" ./tests/e2e/upgrade/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-12-22-165425 True False 53m Cluster version is 4.15.0-0.nightly-2023-12-22-165425' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-12-22-165425 True False 53m Cluster version is 4.15.0-0.nightly-2023-12-22-165425' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/upgrade/render.sh ++ export SUITE_DIR=./tests/e2e/upgrade ++ SUITE_DIR=./tests/e2e/upgrade ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/upgrade ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + export JAEGER_NAME + '[' true = true ']' + skip_test upgrade 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade + warning 'upgrade: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade: Test not supported in OpenShift\e[0m' WAR: upgrade: Test not supported in OpenShift + '[' true = true ']' + skip_test upgrade-from-latest-release 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade-from-latest-release + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade-from-latest-release + warning 'upgrade-from-latest-release: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade-from-latest-release: Test not supported in OpenShift\e[0m' WAR: upgrade-from-latest-release: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running upgrade E2E tests' Running upgrade E2E tests + cd tests/e2e/upgrade/_build + set +e + KUBECONFIG=/tmp/kubeconfig-1441553849 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-w25v5qjw-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 08:19:50 | artifacts | Creating namespace: kuttl-test-settling-jennet logger.go:42: 08:19:50 | artifacts | artifacts events from ns kuttl-test-settling-jennet: logger.go:42: 08:19:50 | artifacts | Deleting namespace: kuttl-test-settling-jennet === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.37s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.20s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name upgrade --report --output /logs/artifacts/upgrade.xml ./artifacts/kuttl-report.xml time="2023-12-25T08:19:56Z" level=debug msg="Setting a new name for the test suites" time="2023-12-25T08:19:56Z" level=debug msg="Removing 'artifacts' TestCase" time="2023-12-25T08:19:56Z" level=debug msg="normalizing test case names" time="2023-12-25T08:19:56Z" level=debug msg="upgrade/artifacts -> upgrade_artifacts" +-------------------+--------+ | NAME | RESULT | +-------------------+--------+ | upgrade_artifacts | passed | +-------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests'