% Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 3831 100 3831 0 0 31687 0 --:--:-- --:--:-- --:--:-- 31925 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 1953 100 1953 0 0 18757 0 --:--:-- --:--:-- --:--:-- 18778 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 350 100 350 0 0 3101 0 --:--:-- --:--:-- --:--:-- 3125 Installing kuttl Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/kubectl-kuttl https://github.com/kudobuilder/kuttl/releases/download/v0.15.0/kubectl-kuttl_0.15.0_linux_x86_64 KUBECONFIG file is: /tmp/kubeconfig-3521674980 for suite in elasticsearch examples generate miscellaneous sidecar streaming ui upgrade; do \ make run-e2e-tests-$suite ; \ done make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh elasticsearch false true + '[' 3 -ne 3 ']' + test_suite_name=elasticsearch + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/elasticsearch.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-elasticsearch make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true \ KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ ./tests/e2e/elasticsearch/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-06-062415 True False 6m21s Cluster version is 4.15.0-0.nightly-2024-01-06-062415' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-06-062415 True False 6m21s Cluster version is 4.15.0-0.nightly-2024-01-06-062415' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/elasticsearch/render.sh ++ export SUITE_DIR=./tests/e2e/elasticsearch ++ SUITE_DIR=./tests/e2e/elasticsearch ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/elasticsearch ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + start_test es-from-aio-to-production + '[' 1 -ne 1 ']' + test_name=es-from-aio-to-production + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-from-aio-to-production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-from-aio-to-production\e[0m' Rendering files for test es-from-aio-to-production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-from-aio-to-production + cd es-from-aio-to-production + jaeger_name=my-jaeger + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 03 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=03 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./03-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch.redundancyPolicy="ZeroRedundancy"' ./03-install.yaml + render_smoke_test my-jaeger true 04 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=04 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./04-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test es-increasing-replicas + '[' 1 -ne 1 ']' + test_name=es-increasing-replicas + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-increasing-replicas' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-increasing-replicas\e[0m' Rendering files for test es-increasing-replicas + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-from-aio-to-production + '[' es-from-aio-to-production '!=' _build ']' + cd .. + mkdir -p es-increasing-replicas + cd es-increasing-replicas + jaeger_name=simple-prod + '[' true = true ']' + jaeger_deployment_mode=production_autoprovisioned + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.replicas=2 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.query.replicas=2 ./02-install.yaml + cp ./01-assert.yaml ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.replicas=2 ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .status.readyReplicas=2 ./02-assert.yaml + render_smoke_test simple-prod true 03 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=03 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./03-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + cp ./02-install.yaml ./04-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.elasticsearch.nodeCount=2 ./04-install.yaml + /tmp/jaeger-tests/bin/gomplate -f ./openshift-check-es-nodes.yaml.template -o ./05-check-es-nodes.yaml + '[' true = true ']' + skip_test es-index-cleaner-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-increasing-replicas + '[' es-increasing-replicas '!=' _build ']' + cd .. + rm -rf es-index-cleaner-upstream + warning 'es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_index_cleaner -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-index-cleaner-autoprov + '[' 1 -ne 1 ']' + test_name=es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-index-cleaner-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-index-cleaner-autoprov\e[0m' Rendering files for test es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-index-cleaner-autoprov + cd es-index-cleaner-autoprov + jaeger_name=test-es-index-cleaner-with-prefix + cronjob_name=test-es-index-cleaner-with-prefix-es-index-cleaner + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + cp ../../es-index-cleaner-upstream/04-assert.yaml ../../es-index-cleaner-upstream/README.md . + render_install_jaeger test-es-index-cleaner-with-prefix production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options.es.index-prefix=""' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.enabled=false ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.numberOfDays=0 ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.esIndexCleaner.schedule="*/1 * * * *"' ./01-install.yaml + render_report_spans test-es-index-cleaner-with-prefix true 5 00 true 02 + '[' 6 -ne 6 ']' + jaeger=test-es-index-cleaner-with-prefix + is_secured=true + number_of_spans=5 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + export JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=5 + DAYS=5 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + sed 's~enabled: false~enabled: true~gi' ./01-install.yaml + CRONJOB_NAME=test-es-index-cleaner-with-prefix-es-index-cleaner + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./04-wait-es-index-cleaner.yaml + /tmp/jaeger-tests/bin/gomplate -f ./01-install.yaml -o ./05-install.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 00 06 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=00 + test_step=06 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=test-es-index-cleaner-with-prefix-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=test-es-index-cleaner-with-prefix-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./06-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./06-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.1"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.1"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.1"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.1 ++ version_ge 5.8.1 5.4 +++ echo 5.8.1 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.1 == 5.8.1 + '[' -n '' ']' + skip_test es-index-cleaner-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-index-cleaner-autoprov + '[' es-index-cleaner-autoprov '!=' _build ']' + cd .. + rm -rf es-index-cleaner-managed + warning 'es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + start_test es-multiinstance + '[' 1 -ne 1 ']' + test_name=es-multiinstance + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-multiinstance' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-multiinstance\e[0m' Rendering files for test es-multiinstance + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-multiinstance + cd es-multiinstance + jaeger_name=instance-1 + render_install_jaeger instance-1 production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=instance-1 + JAEGER_NAME=instance-1 + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f ./03-create-second-instance.yaml.template -o 03-create-second-instance.yaml + '[' true = true ']' + skip_test es-rollover-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-rollover-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-multiinstance + '[' es-multiinstance '!=' _build ']' + cd .. + rm -rf es-rollover-upstream + warning 'es-rollover-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_rollover -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-rollover-autoprov + '[' 1 -ne 1 ']' + test_name=es-rollover-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-rollover-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-rollover-autoprov\e[0m' Rendering files for test es-rollover-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-rollover-autoprov + cd es-rollover-autoprov + cp ../../es-rollover-upstream/05-assert.yaml ../../es-rollover-upstream/05-install.yaml ../../es-rollover-upstream/README.md . + jaeger_name=my-jaeger + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_report_spans my-jaeger true 2 00 true 02 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 00 03 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=00 + test_step=03 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./03-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./03-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 01 04 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=01 + test_step=04 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=01 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./04-check-indices.yaml + JOB_NUMBER=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./04-assert.yaml + render_report_spans my-jaeger true 2 02 true 06 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=02 + ensure_reported_spans=true + test_step=06 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=02 + JOB_NUMBER=02 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./06-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./06-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 02 07 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=02 + test_step=07 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=02 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./07-check-indices.yaml + JOB_NUMBER=02 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./07-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' 03 08 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + job_number=03 + test_step=08 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=03 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./08-check-indices.yaml + JOB_NUMBER=03 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./08-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' 04 09 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + job_number=04 + test_step=09 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=04 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./09-check-indices.yaml + JOB_NUMBER=04 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./09-assert.yaml + render_report_spans my-jaeger true 2 03 true 10 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=03 + ensure_reported_spans=true + test_step=10 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=03 + JOB_NUMBER=03 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./10-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./10-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + CRONJOB_NAME=my-jaeger-es-rollover + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./11-wait-rollover.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-000002'\'',' 05 11 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-000002'\'',' + job_number=05 + test_step=11 + escape_command ''\''--name'\'', '\''jaeger-span-000002'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-000002'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-000002'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-000002'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=05 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./11-check-indices.yaml + JOB_NUMBER=05 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./11-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' 06 12 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + job_number=06 + test_step=12 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=06 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./12-check-indices.yaml + JOB_NUMBER=06 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./12-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.1"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.1"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.1"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.1 ++ version_ge 5.8.1 5.4 +++ echo 5.8.1 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.1 == 5.8.1 + '[' -n '' ']' + skip_test es-rollover-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-rollover-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-rollover-autoprov + '[' es-rollover-autoprov '!=' _build ']' + cd .. + rm -rf es-rollover-managed + warning 'es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + skip_test es-spark-dependencies 'This test is not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=es-spark-dependencies + message='This test is not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + rm -rf es-spark-dependencies + warning 'es-spark-dependencies: This test is not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-spark-dependencies: This test is not supported in OpenShift\e[0m' WAR: es-spark-dependencies: This test is not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running elasticsearch E2E tests' Running elasticsearch E2E tests + cd tests/e2e/elasticsearch/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3521674980 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/es-from-aio-to-production === PAUSE kuttl/harness/es-from-aio-to-production === RUN kuttl/harness/es-increasing-replicas === PAUSE kuttl/harness/es-increasing-replicas === RUN kuttl/harness/es-index-cleaner-autoprov === PAUSE kuttl/harness/es-index-cleaner-autoprov === RUN kuttl/harness/es-multiinstance === PAUSE kuttl/harness/es-multiinstance === RUN kuttl/harness/es-rollover-autoprov === PAUSE kuttl/harness/es-rollover-autoprov === RUN kuttl/harness/es-simple-prod === PAUSE kuttl/harness/es-simple-prod === CONT kuttl/harness/artifacts logger.go:42: 06:58:21 | artifacts | Creating namespace: kuttl-test-new-fawn logger.go:42: 06:58:21 | artifacts | artifacts events from ns kuttl-test-new-fawn: logger.go:42: 06:58:21 | artifacts | Deleting namespace: kuttl-test-new-fawn === CONT kuttl/harness/es-multiinstance logger.go:42: 06:58:28 | es-multiinstance | Ignoring 03-create-second-instance.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 06:58:28 | es-multiinstance | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 06:58:28 | es-multiinstance | Creating namespace: kuttl-test-vast-rattler logger.go:42: 06:58:28 | es-multiinstance/0-clear-namespace | starting test step 0-clear-namespace logger.go:42: 06:58:28 | es-multiinstance/0-clear-namespace | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --ignore-not-found=true] logger.go:42: 06:58:28 | es-multiinstance/0-clear-namespace | test step completed 0-clear-namespace logger.go:42: 06:58:28 | es-multiinstance/1-install | starting test step 1-install logger.go:42: 06:58:28 | es-multiinstance/1-install | Jaeger:kuttl-test-vast-rattler/instance-1 created logger.go:42: 06:59:09 | es-multiinstance/1-install | test step completed 1-install logger.go:42: 06:59:09 | es-multiinstance/2-create-namespace | starting test step 2-create-namespace logger.go:42: 06:59:09 | es-multiinstance/2-create-namespace | running command: [sh -c kubectl create namespace jaeger-e2e-multiinstance-test] logger.go:42: 06:59:10 | es-multiinstance/2-create-namespace | namespace/jaeger-e2e-multiinstance-test created logger.go:42: 06:59:10 | es-multiinstance/2-create-namespace | test step completed 2-create-namespace logger.go:42: 06:59:10 | es-multiinstance/3-create-second-instance | starting test step 3-create-second-instance logger.go:42: 06:59:10 | es-multiinstance/3-create-second-instance | running command: [sh -c kubectl apply -f ./01-install.yaml -n jaeger-e2e-multiinstance-test] logger.go:42: 06:59:12 | es-multiinstance/3-create-second-instance | jaeger.jaegertracing.io/instance-1 created logger.go:42: 06:59:12 | es-multiinstance/3-create-second-instance | running command: [sh -c /tmp/jaeger-tests/bin/kubectl-kuttl assert ./01-assert.yaml -n jaeger-e2e-multiinstance-test --timeout 1000] logger.go:42: 07:00:03 | es-multiinstance/3-create-second-instance | assert is valid logger.go:42: 07:00:03 | es-multiinstance/3-create-second-instance | test step completed 3-create-second-instance logger.go:42: 07:00:03 | es-multiinstance/4-check-secrets | starting test step 4-check-secrets logger.go:42: 07:00:03 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n $NAMESPACE > secret1] logger.go:42: 07:00:04 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n jaeger-e2e-multiinstance-test > secret2] logger.go:42: 07:00:04 | es-multiinstance/4-check-secrets | running command: [sh -c cmp --silent secret1 secret2 || exit 0] logger.go:42: 07:00:04 | es-multiinstance/4-check-secrets | test step completed 4-check-secrets logger.go:42: 07:00:04 | es-multiinstance/5-delete | starting test step 5-delete logger.go:42: 07:00:04 | es-multiinstance/5-delete | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --wait=false] logger.go:42: 07:00:04 | es-multiinstance/5-delete | namespace "jaeger-e2e-multiinstance-test" deleted logger.go:42: 07:00:04 | es-multiinstance/5-delete | test step completed 5-delete logger.go:42: 07:00:04 | es-multiinstance | es-multiinstance events from ns kuttl-test-vast-rattler: logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestvastrattlerinstance1-1-598c6dd62mbfr Binding Scheduled Successfully assigned kuttl-test-vast-rattler/elasticsearch-cdm-kuttltestvastrattlerinstance1-1-598c6dd62mbfr to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestvastrattlerinstance1-1-598c6dd62mbfr AddedInterface Add eth0 [10.129.2.17/23] from ovn-kubernetes logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestvastrattlerinstance1-1-598c6dd62mbfr.spec.containers{elasticsearch} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:37 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestvastrattlerinstance1-1-598c6dd697 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestvastrattlerinstance1-1-598c6dd62mbfr replicaset-controller logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:37 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestvastrattlerinstance1-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestvastrattlerinstance1-1-598c6dd697 to 1 deployment-controller logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:37 +0000 UTC Normal Pod instance-1-collector-8bd877f76-k9whh Binding Scheduled Successfully assigned kuttl-test-vast-rattler/instance-1-collector-8bd877f76-k9whh to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:37 +0000 UTC Normal Pod instance-1-collector-8bd877f76-k9whh AddedInterface Add eth0 [10.128.2.19/23] from ovn-kubernetes logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:37 +0000 UTC Normal Pod instance-1-collector-8bd877f76-k9whh.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:37 +0000 UTC Normal ReplicaSet.apps instance-1-collector-8bd877f76 SuccessfulCreate Created pod: instance-1-collector-8bd877f76-k9whh replicaset-controller logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:37 +0000 UTC Normal Deployment.apps instance-1-collector ScalingReplicaSet Scaled up replica set instance-1-collector-8bd877f76 to 1 deployment-controller logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:37 +0000 UTC Normal Pod instance-1-query-755b896f84-qqvpb Binding Scheduled Successfully assigned kuttl-test-vast-rattler/instance-1-query-755b896f84-qqvpb to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:37 +0000 UTC Normal ReplicaSet.apps instance-1-query-755b896f84 SuccessfulCreate Created pod: instance-1-query-755b896f84-qqvpb replicaset-controller logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:37 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-755b896f84 to 1 deployment-controller logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:38 +0000 UTC Normal Pod instance-1-query-755b896f84-qqvpb AddedInterface Add eth0 [10.131.0.21/23] from ovn-kubernetes logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:38 +0000 UTC Normal Pod instance-1-query-755b896f84-qqvpb.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:42 +0000 UTC Normal Pod instance-1-collector-8bd877f76-k9whh.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" in 4.868s (4.868s including waiting) kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:42 +0000 UTC Normal Pod instance-1-collector-8bd877f76-k9whh.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:42 +0000 UTC Normal Pod instance-1-collector-8bd877f76-k9whh.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:42 +0000 UTC Normal Pod instance-1-query-755b896f84-qqvpb.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" in 4.887s (4.887s including waiting) kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestvastrattlerinstance1-1-598c6dd62mbfr.spec.containers{elasticsearch} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" in 5.463s (5.463s including waiting) kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestvastrattlerinstance1-1-598c6dd62mbfr.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestvastrattlerinstance1-1-598c6dd62mbfr.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestvastrattlerinstance1-1-598c6dd62mbfr.spec.containers{proxy} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:43 +0000 UTC Warning Pod instance-1-collector-8bd877f76-k9whh.spec.containers{jaeger-collector} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:43 +0000 UTC Normal Pod instance-1-query-755b896f84-qqvpb.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:43 +0000 UTC Normal Pod instance-1-query-755b896f84-qqvpb.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:43 +0000 UTC Normal Pod instance-1-query-755b896f84-qqvpb.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestvastrattlerinstance1-1-598c6dd62mbfr.spec.containers{proxy} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" in 3.15s (3.15s including waiting) kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestvastrattlerinstance1-1-598c6dd62mbfr.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestvastrattlerinstance1-1-598c6dd62mbfr.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:48 +0000 UTC Normal Pod instance-1-collector-8bd877f76-k9whh.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:48 +0000 UTC Normal Pod instance-1-query-755b896f84-qqvpb.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" in 5s (5s including waiting) kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:48 +0000 UTC Normal Pod instance-1-query-755b896f84-qqvpb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:48 +0000 UTC Normal Pod instance-1-query-755b896f84-qqvpb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:48 +0000 UTC Normal Pod instance-1-query-755b896f84-qqvpb.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:49 +0000 UTC Normal Pod instance-1-query-755b896f84-qqvpb.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" in 1.612s (1.612s including waiting) kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:49 +0000 UTC Normal Pod instance-1-query-755b896f84-qqvpb.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:49 +0000 UTC Normal Pod instance-1-query-755b896f84-qqvpb.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:50 +0000 UTC Normal Pod instance-1-query-755b896f84-qqvpb.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:51 +0000 UTC Warning Pod instance-1-query-755b896f84-qqvpb.spec.containers{jaeger-query} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:54 +0000 UTC Warning Pod instance-1-collector-8bd877f76-k9whh.spec.containers{jaeger-collector} BackOff Back-off restarting failed container jaeger-collector in pod instance-1-collector-8bd877f76-k9whh_kuttl-test-vast-rattler(a2781085-77d2-49ce-b70f-f524313e37b3) kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:56 +0000 UTC Warning Pod instance-1-query-755b896f84-qqvpb.spec.containers{jaeger-query} BackOff Back-off restarting failed container jaeger-query in pod instance-1-query-755b896f84-qqvpb_kuttl-test-vast-rattler(5c8ca645-def0-44f2-ab84-975bdcaac0d6) kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:58:57 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestvastrattlerinstance1-1-598c6dd62mbfr.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:18 +0000 UTC Normal Pod instance-1-query-755b896f84-qqvpb.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:18 +0000 UTC Normal Pod instance-1-query-755b896f84-qqvpb.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:18 +0000 UTC Normal Pod instance-1-query-755b896f84-qqvpb.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:18 +0000 UTC Normal ReplicaSet.apps instance-1-query-755b896f84 SuccessfulDelete Deleted pod: instance-1-query-755b896f84-qqvpb replicaset-controller logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:18 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled down replica set instance-1-query-755b896f84 to 0 from 1 deployment-controller logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:19 +0000 UTC Normal Pod instance-1-query-64bff58ddd-fkkhh Binding Scheduled Successfully assigned kuttl-test-vast-rattler/instance-1-query-64bff58ddd-fkkhh to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:19 +0000 UTC Normal Pod instance-1-query-64bff58ddd-fkkhh AddedInterface Add eth0 [10.131.0.22/23] from ovn-kubernetes logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:19 +0000 UTC Normal Pod instance-1-query-64bff58ddd-fkkhh.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:19 +0000 UTC Normal Pod instance-1-query-64bff58ddd-fkkhh.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:19 +0000 UTC Normal Pod instance-1-query-64bff58ddd-fkkhh.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:19 +0000 UTC Normal Pod instance-1-query-64bff58ddd-fkkhh.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:19 +0000 UTC Normal Pod instance-1-query-64bff58ddd-fkkhh.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:19 +0000 UTC Normal Pod instance-1-query-64bff58ddd-fkkhh.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:19 +0000 UTC Normal Pod instance-1-query-64bff58ddd-fkkhh.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:19 +0000 UTC Normal ReplicaSet.apps instance-1-query-64bff58ddd SuccessfulCreate Created pod: instance-1-query-64bff58ddd-fkkhh replicaset-controller logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:19 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-64bff58ddd to 1 deployment-controller logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:20 +0000 UTC Normal Pod instance-1-query-64bff58ddd-fkkhh.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:20 +0000 UTC Normal Pod instance-1-query-64bff58ddd-fkkhh.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:26 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:26 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:56 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:56 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:58 +0000 UTC Normal Pod instance-1-query-64bff58ddd-fkkhh.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:58 +0000 UTC Normal Pod instance-1-query-64bff58ddd-fkkhh.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:58 +0000 UTC Normal Pod instance-1-query-64bff58ddd-fkkhh.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:58 +0000 UTC Normal ReplicaSet.apps instance-1-query-64bff58ddd SuccessfulDelete Deleted pod: instance-1-query-64bff58ddd-fkkhh replicaset-controller logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:58 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled down replica set instance-1-query-64bff58ddd to 0 from 1 deployment-controller logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:59 +0000 UTC Normal Pod instance-1-query-755b896f84-hpdtv Binding Scheduled Successfully assigned kuttl-test-vast-rattler/instance-1-query-755b896f84-hpdtv to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:59 +0000 UTC Normal Pod instance-1-query-755b896f84-hpdtv AddedInterface Add eth0 [10.131.0.25/23] from ovn-kubernetes logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:59 +0000 UTC Normal Pod instance-1-query-755b896f84-hpdtv.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:59 +0000 UTC Normal ReplicaSet.apps instance-1-query-755b896f84 SuccessfulCreate Created pod: instance-1-query-755b896f84-hpdtv replicaset-controller logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 06:59:59 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-755b896f84 to 1 from 0 deployment-controller logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 07:00:00 +0000 UTC Normal Pod instance-1-query-755b896f84-hpdtv.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 07:00:00 +0000 UTC Normal Pod instance-1-query-755b896f84-hpdtv.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 07:00:00 +0000 UTC Normal Pod instance-1-query-755b896f84-hpdtv.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 07:00:00 +0000 UTC Normal Pod instance-1-query-755b896f84-hpdtv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 07:00:00 +0000 UTC Normal Pod instance-1-query-755b896f84-hpdtv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 07:00:00 +0000 UTC Normal Pod instance-1-query-755b896f84-hpdtv.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 07:00:00 +0000 UTC Normal Pod instance-1-query-755b896f84-hpdtv.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:00:04 | es-multiinstance | 2024-01-08 07:00:00 +0000 UTC Normal Pod instance-1-query-755b896f84-hpdtv.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:00:04 | es-multiinstance | Deleting namespace: kuttl-test-vast-rattler === CONT kuttl/harness/es-simple-prod logger.go:42: 07:00:12 | es-simple-prod | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:00:12 | es-simple-prod | Creating namespace: kuttl-test-stunning-eft logger.go:42: 07:00:12 | es-simple-prod | es-simple-prod events from ns kuttl-test-stunning-eft: logger.go:42: 07:00:12 | es-simple-prod | Deleting namespace: kuttl-test-stunning-eft === CONT kuttl/harness/es-rollover-autoprov logger.go:42: 07:00:18 | es-rollover-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:00:18 | es-rollover-autoprov | Creating namespace: kuttl-test-relevant-scorpion logger.go:42: 07:00:18 | es-rollover-autoprov/1-install | starting test step 1-install logger.go:42: 07:00:18 | es-rollover-autoprov/1-install | Jaeger:kuttl-test-relevant-scorpion/my-jaeger created logger.go:42: 07:00:54 | es-rollover-autoprov/1-install | test step completed 1-install logger.go:42: 07:00:54 | es-rollover-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:00:54 | es-rollover-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:00:56 | es-rollover-autoprov/2-report-spans | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:01:03 | es-rollover-autoprov/2-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:01:04 | es-rollover-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:01:04 | es-rollover-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:01:31 | es-rollover-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 07:01:31 | es-rollover-autoprov/3-check-indices | starting test step 3-check-indices logger.go:42: 07:01:31 | es-rollover-autoprov/3-check-indices | Job:kuttl-test-relevant-scorpion/00-check-indices created logger.go:42: 07:01:35 | es-rollover-autoprov/3-check-indices | test step completed 3-check-indices logger.go:42: 07:01:35 | es-rollover-autoprov/4-check-indices | starting test step 4-check-indices logger.go:42: 07:01:35 | es-rollover-autoprov/4-check-indices | Job:kuttl-test-relevant-scorpion/01-check-indices created logger.go:42: 07:01:39 | es-rollover-autoprov/4-check-indices | test step completed 4-check-indices logger.go:42: 07:01:39 | es-rollover-autoprov/5-install | starting test step 5-install logger.go:42: 07:01:39 | es-rollover-autoprov/5-install | Jaeger:kuttl-test-relevant-scorpion/my-jaeger updated logger.go:42: 07:01:46 | es-rollover-autoprov/5-install | test step completed 5-install logger.go:42: 07:01:46 | es-rollover-autoprov/6-report-spans | starting test step 6-report-spans logger.go:42: 07:01:46 | es-rollover-autoprov/6-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:01:55 | es-rollover-autoprov/6-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JOB_NUMBER=02 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-02-job.yaml] logger.go:42: 07:01:55 | es-rollover-autoprov/6-report-spans | running command: [sh -c kubectl apply -f report-span-02-job.yaml -n $NAMESPACE] logger.go:42: 07:01:55 | es-rollover-autoprov/6-report-spans | job.batch/02-report-span created logger.go:42: 07:02:19 | es-rollover-autoprov/6-report-spans | test step completed 6-report-spans logger.go:42: 07:02:19 | es-rollover-autoprov/7-check-indices | starting test step 7-check-indices logger.go:42: 07:02:19 | es-rollover-autoprov/7-check-indices | Job:kuttl-test-relevant-scorpion/02-check-indices created logger.go:42: 07:02:23 | es-rollover-autoprov/7-check-indices | test step completed 7-check-indices logger.go:42: 07:02:23 | es-rollover-autoprov/8-check-indices | starting test step 8-check-indices logger.go:42: 07:02:23 | es-rollover-autoprov/8-check-indices | Job:kuttl-test-relevant-scorpion/03-check-indices created logger.go:42: 07:02:27 | es-rollover-autoprov/8-check-indices | test step completed 8-check-indices logger.go:42: 07:02:27 | es-rollover-autoprov/9-check-indices | starting test step 9-check-indices logger.go:42: 07:02:27 | es-rollover-autoprov/9-check-indices | Job:kuttl-test-relevant-scorpion/04-check-indices created logger.go:42: 07:02:30 | es-rollover-autoprov/9-check-indices | test step completed 9-check-indices logger.go:42: 07:02:30 | es-rollover-autoprov/10-report-spans | starting test step 10-report-spans logger.go:42: 07:02:30 | es-rollover-autoprov/10-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:02:38 | es-rollover-autoprov/10-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JOB_NUMBER=03 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-03-job.yaml] logger.go:42: 07:02:39 | es-rollover-autoprov/10-report-spans | running command: [sh -c kubectl apply -f report-span-03-job.yaml -n $NAMESPACE] logger.go:42: 07:02:39 | es-rollover-autoprov/10-report-spans | job.batch/03-report-span created logger.go:42: 07:03:03 | es-rollover-autoprov/10-report-spans | test step completed 10-report-spans logger.go:42: 07:03:03 | es-rollover-autoprov/11-check-indices | starting test step 11-check-indices logger.go:42: 07:03:03 | es-rollover-autoprov/11-check-indices | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob my-jaeger-es-rollover --namespace $NAMESPACE] logger.go:42: 07:03:13 | es-rollover-autoprov/11-check-indices | time="2024-01-08T07:03:13Z" level=debug msg="Checking if the my-jaeger-es-rollover CronJob exists" logger.go:42: 07:03:13 | es-rollover-autoprov/11-check-indices | time="2024-01-08T07:03:13Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 07:03:13 | es-rollover-autoprov/11-check-indices | time="2024-01-08T07:03:13Z" level=info msg="Cronjob my-jaeger-es-rollover found successfully" logger.go:42: 07:03:13 | es-rollover-autoprov/11-check-indices | time="2024-01-08T07:03:13Z" level=debug msg="Waiting for the next scheduled job from my-jaeger-es-rollover cronjob" logger.go:42: 07:03:13 | es-rollover-autoprov/11-check-indices | time="2024-01-08T07:03:13Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:03:23 | es-rollover-autoprov/11-check-indices | time="2024-01-08T07:03:23Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:03:33 | es-rollover-autoprov/11-check-indices | time="2024-01-08T07:03:33Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:03:43 | es-rollover-autoprov/11-check-indices | time="2024-01-08T07:03:43Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:03:53 | es-rollover-autoprov/11-check-indices | time="2024-01-08T07:03:53Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:04:03 | es-rollover-autoprov/11-check-indices | time="2024-01-08T07:04:03Z" level=info msg="Job of owner my-jaeger-es-rollover succeeded after my-jaeger-es-rollover 50.052281237s" logger.go:42: 07:04:03 | es-rollover-autoprov/11-check-indices | Job:kuttl-test-relevant-scorpion/05-check-indices created logger.go:42: 07:04:07 | es-rollover-autoprov/11-check-indices | test step completed 11-check-indices logger.go:42: 07:04:07 | es-rollover-autoprov/12-check-indices | starting test step 12-check-indices logger.go:42: 07:04:07 | es-rollover-autoprov/12-check-indices | Job:kuttl-test-relevant-scorpion/06-check-indices created logger.go:42: 07:04:12 | es-rollover-autoprov/12-check-indices | test step completed 12-check-indices logger.go:42: 07:04:12 | es-rollover-autoprov | es-rollover-autoprov events from ns kuttl-test-relevant-scorpion: logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:25 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestrelevantscorpionmyjaeger-1-585bbcbfcc SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestrelevantscorpionmyjaeger-1-585bdsm42 replicaset-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelevantscorpionmyjaeger-1-585bdsm42 Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/elasticsearch-cdm-kuttltestrelevantscorpionmyjaeger-1-585bdsm42 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:25 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestrelevantscorpionmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestrelevantscorpionmyjaeger-1-585bbcbfcc to 1 deployment-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelevantscorpionmyjaeger-1-585bdsm42 AddedInterface Add eth0 [10.129.2.19/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelevantscorpionmyjaeger-1-585bdsm42.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelevantscorpionmyjaeger-1-585bdsm42.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelevantscorpionmyjaeger-1-585bdsm42.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelevantscorpionmyjaeger-1-585bdsm42.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelevantscorpionmyjaeger-1-585bdsm42.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelevantscorpionmyjaeger-1-585bdsm42.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:36 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestrelevantscorpionmyjaeger-1-585bdsm42.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:41 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestrelevantscorpionmyjaeger-1-585bdsm42.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:52 +0000 UTC Normal Pod my-jaeger-collector-5b59c747db-2h49f Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/my-jaeger-collector-5b59c747db-2h49f to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:52 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5b59c747db SuccessfulCreate Created pod: my-jaeger-collector-5b59c747db-2h49f replicaset-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:52 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-5b59c747db to 1 deployment-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:52 +0000 UTC Normal Pod my-jaeger-query-d884b9cf-qjnl5 Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/my-jaeger-query-d884b9cf-qjnl5 to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:52 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-d884b9cf SuccessfulCreate Created pod: my-jaeger-query-d884b9cf-qjnl5 replicaset-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:52 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-d884b9cf to 1 deployment-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:53 +0000 UTC Normal Pod my-jaeger-collector-5b59c747db-2h49f AddedInterface Add eth0 [10.128.2.21/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:53 +0000 UTC Normal Pod my-jaeger-collector-5b59c747db-2h49f.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:53 +0000 UTC Normal Pod my-jaeger-collector-5b59c747db-2h49f.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:53 +0000 UTC Normal Pod my-jaeger-collector-5b59c747db-2h49f.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:53 +0000 UTC Normal Pod my-jaeger-query-d884b9cf-qjnl5 AddedInterface Add eth0 [10.131.0.26/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:53 +0000 UTC Normal Pod my-jaeger-query-d884b9cf-qjnl5.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:53 +0000 UTC Normal Pod my-jaeger-query-d884b9cf-qjnl5.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:53 +0000 UTC Normal Pod my-jaeger-query-d884b9cf-qjnl5.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:53 +0000 UTC Normal Pod my-jaeger-query-d884b9cf-qjnl5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:53 +0000 UTC Normal Pod my-jaeger-query-d884b9cf-qjnl5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:53 +0000 UTC Normal Pod my-jaeger-query-d884b9cf-qjnl5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:53 +0000 UTC Normal Pod my-jaeger-query-d884b9cf-qjnl5.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:53 +0000 UTC Normal Pod my-jaeger-query-d884b9cf-qjnl5.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:00:53 +0000 UTC Normal Pod my-jaeger-query-d884b9cf-qjnl5.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:00 +0000 UTC Normal Pod my-jaeger-query-d884b9cf-qjnl5.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:00 +0000 UTC Normal Pod my-jaeger-query-d884b9cf-qjnl5.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:00 +0000 UTC Normal Pod my-jaeger-query-d884b9cf-qjnl5.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:00 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-d884b9cf SuccessfulDelete Deleted pod: my-jaeger-query-d884b9cf-qjnl5 replicaset-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:00 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-d884b9cf to 0 from 1 deployment-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:01 +0000 UTC Normal Pod my-jaeger-query-7db65c8854-vpv96 Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/my-jaeger-query-7db65c8854-vpv96 to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:01 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7db65c8854 SuccessfulCreate Created pod: my-jaeger-query-7db65c8854-vpv96 replicaset-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:01 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7db65c8854 to 1 deployment-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:02 +0000 UTC Normal Pod my-jaeger-query-7db65c8854-vpv96 AddedInterface Add eth0 [10.131.0.27/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:02 +0000 UTC Normal Pod my-jaeger-query-7db65c8854-vpv96.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:02 +0000 UTC Normal Pod my-jaeger-query-7db65c8854-vpv96.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:02 +0000 UTC Normal Pod my-jaeger-query-7db65c8854-vpv96.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:02 +0000 UTC Normal Pod my-jaeger-query-7db65c8854-vpv96.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:02 +0000 UTC Normal Pod my-jaeger-query-7db65c8854-vpv96.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:02 +0000 UTC Normal Pod my-jaeger-query-7db65c8854-vpv96.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:02 +0000 UTC Normal Pod my-jaeger-query-7db65c8854-vpv96.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:02 +0000 UTC Normal Pod my-jaeger-query-7db65c8854-vpv96.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:02 +0000 UTC Normal Pod my-jaeger-query-7db65c8854-vpv96.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:04 +0000 UTC Normal Pod 00-report-span-jqw2r Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/00-report-span-jqw2r to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:04 +0000 UTC Normal Pod 00-report-span-jqw2r AddedInterface Add eth0 [10.128.2.22/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:04 +0000 UTC Normal Pod 00-report-span-jqw2r.spec.containers{asserts-container} Pulling Pulling image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:04 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-jqw2r job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:08 +0000 UTC Normal Pod 00-report-span-jqw2r.spec.containers{asserts-container} Pulled Successfully pulled image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" in 3.236s (3.236s including waiting) kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:08 +0000 UTC Normal Pod 00-report-span-jqw2r.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:08 +0000 UTC Normal Pod 00-report-span-jqw2r.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:09 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:09 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:09 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:30 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:31 +0000 UTC Normal Pod 00-check-indices-2vctv Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/00-check-indices-2vctv to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:31 +0000 UTC Normal Pod 00-check-indices-2vctv AddedInterface Add eth0 [10.128.2.23/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:31 +0000 UTC Normal Pod 00-check-indices-2vctv.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:31 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-2vctv job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:32 +0000 UTC Normal Pod 00-check-indices-2vctv.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:32 +0000 UTC Normal Pod 00-check-indices-2vctv.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:34 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:35 +0000 UTC Normal Pod 01-check-indices-tbxjg Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/01-check-indices-tbxjg to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:35 +0000 UTC Normal Pod 01-check-indices-tbxjg AddedInterface Add eth0 [10.128.2.24/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:35 +0000 UTC Normal Pod 01-check-indices-tbxjg.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:35 +0000 UTC Normal Pod 01-check-indices-tbxjg.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:35 +0000 UTC Normal Pod 01-check-indices-tbxjg.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:35 +0000 UTC Normal Job.batch 01-check-indices SuccessfulCreate Created pod: 01-check-indices-tbxjg job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:38 +0000 UTC Normal Job.batch 01-check-indices Completed Job completed job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:39 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:39 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:40 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-9x2tr Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/my-jaeger-es-rollover-create-mapping-9x2tr to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:40 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping SuccessfulCreate Created pod: my-jaeger-es-rollover-create-mapping-9x2tr job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:41 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-9x2tr AddedInterface Add eth0 [10.128.2.25/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:41 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-9x2tr.spec.containers{my-jaeger-es-rollover-create-mapping} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:43 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-9x2tr.spec.containers{my-jaeger-es-rollover-create-mapping} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" in 1.799s (1.799s including waiting) kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:43 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-9x2tr.spec.containers{my-jaeger-es-rollover-create-mapping} Created Created container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:43 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-9x2tr.spec.containers{my-jaeger-es-rollover-create-mapping} Started Started container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:45 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping Completed Job completed job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:46 +0000 UTC Normal Pod my-jaeger-collector-5b59c747db-2h49f.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:46 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5b59c747db SuccessfulDelete Deleted pod: my-jaeger-collector-5b59c747db-2h49f replicaset-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:46 +0000 UTC Normal Pod my-jaeger-collector-77696449df-jq7sv Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/my-jaeger-collector-77696449df-jq7sv to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:46 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-77696449df SuccessfulCreate Created pod: my-jaeger-collector-77696449df-jq7sv replicaset-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:46 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-5b59c747db to 0 from 1 deployment-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:46 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-77696449df to 1 deployment-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:46 +0000 UTC Normal Pod my-jaeger-query-7db65c8854-vpv96.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:46 +0000 UTC Normal Pod my-jaeger-query-7db65c8854-vpv96.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:46 +0000 UTC Normal Pod my-jaeger-query-7db65c8854-vpv96.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:46 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7db65c8854 SuccessfulDelete Deleted pod: my-jaeger-query-7db65c8854-vpv96 replicaset-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:46 +0000 UTC Normal Pod my-jaeger-query-bcb94595-76q94 Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/my-jaeger-query-bcb94595-76q94 to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:46 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-bcb94595 SuccessfulCreate Created pod: my-jaeger-query-bcb94595-76q94 replicaset-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:46 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-7db65c8854 to 0 from 1 deployment-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:46 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-bcb94595 to 1 deployment-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:47 +0000 UTC Normal Pod my-jaeger-collector-77696449df-jq7sv AddedInterface Add eth0 [10.128.2.26/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:47 +0000 UTC Normal Pod my-jaeger-collector-77696449df-jq7sv.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:47 +0000 UTC Normal Pod my-jaeger-collector-77696449df-jq7sv.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:47 +0000 UTC Normal Pod my-jaeger-collector-77696449df-jq7sv.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:47 +0000 UTC Normal Pod my-jaeger-query-bcb94595-76q94 AddedInterface Add eth0 [10.131.0.28/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:47 +0000 UTC Normal Pod my-jaeger-query-bcb94595-76q94.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:47 +0000 UTC Normal Pod my-jaeger-query-bcb94595-76q94.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:47 +0000 UTC Normal Pod my-jaeger-query-bcb94595-76q94.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:47 +0000 UTC Normal Pod my-jaeger-query-bcb94595-76q94.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:47 +0000 UTC Normal Pod my-jaeger-query-bcb94595-76q94.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:47 +0000 UTC Normal Pod my-jaeger-query-bcb94595-76q94.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:47 +0000 UTC Normal Pod my-jaeger-query-bcb94595-76q94.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:47 +0000 UTC Normal Pod my-jaeger-query-bcb94595-76q94.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:47 +0000 UTC Normal Pod my-jaeger-query-bcb94595-76q94.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:55 +0000 UTC Normal Pod 02-report-span-tftsx Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/02-report-span-tftsx to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:55 +0000 UTC Normal Job.batch 02-report-span SuccessfulCreate Created pod: 02-report-span-tftsx job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:56 +0000 UTC Normal Pod 02-report-span-tftsx AddedInterface Add eth0 [10.128.2.27/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:56 +0000 UTC Normal Pod 02-report-span-tftsx.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:56 +0000 UTC Normal Pod 02-report-span-tftsx.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:01:56 +0000 UTC Normal Pod 02-report-span-tftsx.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28411622-m85t8 Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/my-jaeger-es-lookback-28411622-m85t8 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28411622-m85t8 AddedInterface Add eth0 [10.128.2.28/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28411622-m85t8.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28411622-m85t8.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28411622-m85t8.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28411622 SuccessfulCreate Created pod: my-jaeger-es-lookback-28411622-m85t8 job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28411622 cronjob-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28411622-5dvrr Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/my-jaeger-es-rollover-28411622-5dvrr to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28411622-5dvrr AddedInterface Add eth0 [10.128.2.29/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28411622-5dvrr.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28411622-5dvrr.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28411622-5dvrr.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28411622 SuccessfulCreate Created pod: my-jaeger-es-rollover-28411622-5dvrr job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28411622 cronjob-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28411622 Completed Job completed job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28411622, status: Complete cronjob-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28411622 Completed Job completed job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28411622, status: Complete cronjob-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:18 +0000 UTC Normal Job.batch 02-report-span Completed Job completed job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:19 +0000 UTC Normal Pod 02-check-indices-2f9bb Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/02-check-indices-2f9bb to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:19 +0000 UTC Normal Pod 02-check-indices-2f9bb AddedInterface Add eth0 [10.128.2.30/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:19 +0000 UTC Normal Pod 02-check-indices-2f9bb.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:19 +0000 UTC Normal Pod 02-check-indices-2f9bb.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:19 +0000 UTC Normal Pod 02-check-indices-2f9bb.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:19 +0000 UTC Normal Job.batch 02-check-indices SuccessfulCreate Created pod: 02-check-indices-2f9bb job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:22 +0000 UTC Normal Job.batch 02-check-indices Completed Job completed job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:23 +0000 UTC Normal Pod 03-check-indices-gm6mj Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/03-check-indices-gm6mj to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:23 +0000 UTC Normal Job.batch 03-check-indices SuccessfulCreate Created pod: 03-check-indices-gm6mj job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:24 +0000 UTC Normal Pod 03-check-indices-gm6mj AddedInterface Add eth0 [10.128.2.31/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:24 +0000 UTC Normal Pod 03-check-indices-gm6mj.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:24 +0000 UTC Normal Pod 03-check-indices-gm6mj.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:24 +0000 UTC Normal Pod 03-check-indices-gm6mj.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:24 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-77696449df-jq7sv horizontal-pod-autoscaler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:26 +0000 UTC Normal Job.batch 03-check-indices Completed Job completed job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:27 +0000 UTC Normal Pod 04-check-indices-lzsrq Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/04-check-indices-lzsrq to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:27 +0000 UTC Normal Job.batch 04-check-indices SuccessfulCreate Created pod: 04-check-indices-lzsrq job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:28 +0000 UTC Normal Pod 04-check-indices-lzsrq AddedInterface Add eth0 [10.128.2.32/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:28 +0000 UTC Normal Pod 04-check-indices-lzsrq.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:28 +0000 UTC Normal Pod 04-check-indices-lzsrq.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:28 +0000 UTC Normal Pod 04-check-indices-lzsrq.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:30 +0000 UTC Normal Job.batch 04-check-indices Completed Job completed job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:39 +0000 UTC Normal Pod 03-report-span-l8sbx Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/03-report-span-l8sbx to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:39 +0000 UTC Normal Job.batch 03-report-span SuccessfulCreate Created pod: 03-report-span-l8sbx job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:40 +0000 UTC Normal Pod 03-report-span-l8sbx AddedInterface Add eth0 [10.128.2.33/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:40 +0000 UTC Normal Pod 03-report-span-l8sbx.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:40 +0000 UTC Normal Pod 03-report-span-l8sbx.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:02:40 +0000 UTC Normal Pod 03-report-span-l8sbx.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28411623-b6f4w Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/my-jaeger-es-lookback-28411623-b6f4w to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28411623-b6f4w AddedInterface Add eth0 [10.128.2.35/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28411623-b6f4w.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28411623-b6f4w.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28411623-b6f4w.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28411623 SuccessfulCreate Created pod: my-jaeger-es-lookback-28411623-b6f4w job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28411623 cronjob-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28411623-gf8s8 Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/my-jaeger-es-rollover-28411623-gf8s8 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28411623-gf8s8 AddedInterface Add eth0 [10.128.2.34/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28411623-gf8s8.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28411623-gf8s8.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28411623-gf8s8.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28411623 SuccessfulCreate Created pod: my-jaeger-es-rollover-28411623-gf8s8 job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28411623 cronjob-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:02 +0000 UTC Normal Job.batch 03-report-span Completed Job completed job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28411623 Completed Job completed job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28411623, status: Complete cronjob-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:04 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28411623 Completed Job completed job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:03:04 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28411623, status: Complete cronjob-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28411624-hdrsg Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/my-jaeger-es-lookback-28411624-hdrsg to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28411624 SuccessfulCreate Created pod: my-jaeger-es-lookback-28411624-hdrsg job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28411624 cronjob-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28411624-g2ccp Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/my-jaeger-es-rollover-28411624-g2ccp to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28411624-g2ccp AddedInterface Add eth0 [10.128.2.37/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28411624-g2ccp.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28411624 SuccessfulCreate Created pod: my-jaeger-es-rollover-28411624-g2ccp job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28411624 cronjob-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28411624-hdrsg AddedInterface Add eth0 [10.128.2.36/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28411624-hdrsg.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28411624-hdrsg.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28411624-hdrsg.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:01 +0000 UTC Normal Pod my-jaeger-es-rollover-28411624-g2ccp.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:01 +0000 UTC Normal Pod my-jaeger-es-rollover-28411624-g2ccp.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:03 +0000 UTC Normal Pod 05-check-indices-z28hf Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/05-check-indices-z28hf to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:03 +0000 UTC Normal Job.batch 05-check-indices SuccessfulCreate Created pod: 05-check-indices-z28hf job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28411624 Completed Job completed job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28411624, status: Complete cronjob-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:04 +0000 UTC Normal Pod 05-check-indices-z28hf AddedInterface Add eth0 [10.128.2.38/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:04 +0000 UTC Normal Pod 05-check-indices-z28hf.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:04 +0000 UTC Normal Pod 05-check-indices-z28hf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:04 +0000 UTC Normal Pod 05-check-indices-z28hf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:04 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28411624 Completed Job completed job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:04 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28411624, status: Complete cronjob-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:07 +0000 UTC Normal Job.batch 05-check-indices Completed Job completed job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:07 +0000 UTC Normal Pod 06-check-indices-qzdgw Binding Scheduled Successfully assigned kuttl-test-relevant-scorpion/06-check-indices-qzdgw to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:07 +0000 UTC Normal Job.batch 06-check-indices SuccessfulCreate Created pod: 06-check-indices-qzdgw job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:08 +0000 UTC Normal Pod 06-check-indices-qzdgw AddedInterface Add eth0 [10.128.2.39/23] from ovn-kubernetes logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:08 +0000 UTC Normal Pod 06-check-indices-qzdgw.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:08 +0000 UTC Normal Pod 06-check-indices-qzdgw.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:08 +0000 UTC Normal Pod 06-check-indices-qzdgw.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:12 | es-rollover-autoprov | 2024-01-08 07:04:11 +0000 UTC Normal Job.batch 06-check-indices Completed Job completed job-controller logger.go:42: 07:04:12 | es-rollover-autoprov | Deleting namespace: kuttl-test-relevant-scorpion === CONT kuttl/harness/es-increasing-replicas logger.go:42: 07:04:20 | es-increasing-replicas | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:04:20 | es-increasing-replicas | Ignoring check-es-nodes.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:04:20 | es-increasing-replicas | Ignoring openshift-check-es-nodes.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:04:20 | es-increasing-replicas | Creating namespace: kuttl-test-unique-humpback logger.go:42: 07:04:20 | es-increasing-replicas/1-install | starting test step 1-install logger.go:42: 07:04:21 | es-increasing-replicas/1-install | Jaeger:kuttl-test-unique-humpback/simple-prod created logger.go:42: 07:04:57 | es-increasing-replicas/1-install | test step completed 1-install logger.go:42: 07:04:57 | es-increasing-replicas/2-install | starting test step 2-install logger.go:42: 07:04:57 | es-increasing-replicas/2-install | Jaeger:kuttl-test-unique-humpback/simple-prod updated logger.go:42: 07:05:14 | es-increasing-replicas/2-install | test step completed 2-install logger.go:42: 07:05:14 | es-increasing-replicas/3-smoke-test | starting test step 3-smoke-test logger.go:42: 07:05:14 | es-increasing-replicas/3-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:05:15 | es-increasing-replicas/3-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:05:23 | es-increasing-replicas/3-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:05:24 | es-increasing-replicas/3-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:05:24 | es-increasing-replicas/3-smoke-test | job.batch/report-span created logger.go:42: 07:05:24 | es-increasing-replicas/3-smoke-test | job.batch/check-span created logger.go:42: 07:05:30 | es-increasing-replicas/3-smoke-test | test step completed 3-smoke-test logger.go:42: 07:05:30 | es-increasing-replicas/4-install | starting test step 4-install logger.go:42: 07:05:30 | es-increasing-replicas/4-install | Jaeger:kuttl-test-unique-humpback/simple-prod updated logger.go:42: 07:05:30 | es-increasing-replicas/4-install | test step completed 4-install logger.go:42: 07:05:30 | es-increasing-replicas/5-check-es-nodes | starting test step 5-check-es-nodes logger.go:42: 07:05:30 | es-increasing-replicas/5-check-es-nodes | running command: [sh -c ./check-es-nodes.sh $NAMESPACE] logger.go:42: 07:05:30 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:05:31 | es-increasing-replicas/5-check-es-nodes | false logger.go:42: 07:05:31 | es-increasing-replicas/5-check-es-nodes | Error: no matches found logger.go:42: 07:05:36 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:05:36 | es-increasing-replicas/5-check-es-nodes | true logger.go:42: 07:05:36 | es-increasing-replicas/5-check-es-nodes | test step completed 5-check-es-nodes logger.go:42: 07:05:36 | es-increasing-replicas | es-increasing-replicas events from ns kuttl-test-unique-humpback: logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:27 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-1-56dcf5bb4c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-1-56dcnq4lf replicaset-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-1-56dcnq4lf Binding Scheduled Successfully assigned kuttl-test-unique-humpback/elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-1-56dcnq4lf to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:27 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-1-56dcf5bb4c to 1 deployment-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-1-56dcnq4lf AddedInterface Add eth0 [10.129.2.20/23] from ovn-kubernetes logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-1-56dcnq4lf.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-1-56dcnq4lf.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-1-56dcnq4lf.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-1-56dcnq4lf.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-1-56dcnq4lf.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-1-56dcnq4lf.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:38 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-1-56dcnq4lf.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:43 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-1-56dcnq4lf.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:54 +0000 UTC Normal Pod simple-prod-collector-6fcf554844-lgbnf Binding Scheduled Successfully assigned kuttl-test-unique-humpback/simple-prod-collector-6fcf554844-lgbnf to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:54 +0000 UTC Warning Pod simple-prod-collector-6fcf554844-lgbnf FailedMount MountVolume.SetUp failed for volume "simple-prod-collector-tls-config-volume" : secret "simple-prod-collector-headless-tls" not found kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:54 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-6fcf554844 SuccessfulCreate Created pod: simple-prod-collector-6fcf554844-lgbnf replicaset-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:54 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-6fcf554844 to 1 deployment-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:54 +0000 UTC Normal Pod simple-prod-query-d47cf7449-cl95w Binding Scheduled Successfully assigned kuttl-test-unique-humpback/simple-prod-query-d47cf7449-cl95w to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:54 +0000 UTC Normal ReplicaSet.apps simple-prod-query-d47cf7449 SuccessfulCreate Created pod: simple-prod-query-d47cf7449-cl95w replicaset-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:54 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-d47cf7449 to 1 deployment-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:55 +0000 UTC Normal Pod simple-prod-collector-6fcf554844-lgbnf AddedInterface Add eth0 [10.128.2.40/23] from ovn-kubernetes logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:55 +0000 UTC Normal Pod simple-prod-collector-6fcf554844-lgbnf.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:55 +0000 UTC Normal Pod simple-prod-collector-6fcf554844-lgbnf.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:55 +0000 UTC Normal Pod simple-prod-collector-6fcf554844-lgbnf.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:55 +0000 UTC Normal Pod simple-prod-query-d47cf7449-cl95w AddedInterface Add eth0 [10.131.0.29/23] from ovn-kubernetes logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:55 +0000 UTC Normal Pod simple-prod-query-d47cf7449-cl95w.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:55 +0000 UTC Normal Pod simple-prod-query-d47cf7449-cl95w.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:55 +0000 UTC Normal Pod simple-prod-query-d47cf7449-cl95w.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:55 +0000 UTC Normal Pod simple-prod-query-d47cf7449-cl95w.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:55 +0000 UTC Normal Pod simple-prod-query-d47cf7449-cl95w.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:55 +0000 UTC Normal Pod simple-prod-query-d47cf7449-cl95w.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:55 +0000 UTC Normal Pod simple-prod-query-d47cf7449-cl95w.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:55 +0000 UTC Normal Pod simple-prod-query-d47cf7449-cl95w.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:55 +0000 UTC Normal Pod simple-prod-query-d47cf7449-cl95w.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:59 +0000 UTC Normal Pod simple-prod-collector-6fcf554844-7mb82 Binding Scheduled Successfully assigned kuttl-test-unique-humpback/simple-prod-collector-6fcf554844-7mb82 to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:59 +0000 UTC Normal Pod simple-prod-collector-6fcf554844-7mb82 AddedInterface Add eth0 [10.131.0.30/23] from ovn-kubernetes logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:59 +0000 UTC Normal Pod simple-prod-collector-6fcf554844-7mb82.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:59 +0000 UTC Normal Pod simple-prod-collector-6fcf554844-7mb82.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:59 +0000 UTC Normal Pod simple-prod-collector-6fcf554844-7mb82.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:59 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-6fcf554844 SuccessfulCreate Created pod: simple-prod-collector-6fcf554844-7mb82 replicaset-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:59 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-6fcf554844 to 2 from 1 deployment-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:59 +0000 UTC Normal Pod simple-prod-query-d47cf7449-s24r8 Binding Scheduled Successfully assigned kuttl-test-unique-humpback/simple-prod-query-d47cf7449-s24r8 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:59 +0000 UTC Normal ReplicaSet.apps simple-prod-query-d47cf7449 SuccessfulCreate Created pod: simple-prod-query-d47cf7449-s24r8 replicaset-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:04:59 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-d47cf7449 to 2 from 1 deployment-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:00 +0000 UTC Normal Pod simple-prod-query-d47cf7449-s24r8 AddedInterface Add eth0 [10.128.2.41/23] from ovn-kubernetes logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:00 +0000 UTC Normal Pod simple-prod-query-d47cf7449-s24r8.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:04 +0000 UTC Normal Pod simple-prod-query-d47cf7449-s24r8.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" in 4.597s (4.597s including waiting) kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:04 +0000 UTC Normal Pod simple-prod-query-d47cf7449-s24r8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:04 +0000 UTC Normal Pod simple-prod-query-d47cf7449-s24r8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:04 +0000 UTC Normal Pod simple-prod-query-d47cf7449-s24r8.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:09 +0000 UTC Normal Pod simple-prod-query-d47cf7449-s24r8.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" in 5.177s (5.177s including waiting) kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:10 +0000 UTC Normal Pod simple-prod-query-d47cf7449-s24r8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:10 +0000 UTC Normal Pod simple-prod-query-d47cf7449-s24r8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:10 +0000 UTC Normal Pod simple-prod-query-d47cf7449-s24r8.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:12 +0000 UTC Normal Pod simple-prod-query-d47cf7449-s24r8.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" in 2.415s (2.415s including waiting) kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:12 +0000 UTC Normal Pod simple-prod-query-d47cf7449-s24r8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:12 +0000 UTC Normal Pod simple-prod-query-d47cf7449-s24r8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:17 +0000 UTC Normal Pod simple-prod-query-d47cf7449-cl95w.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:17 +0000 UTC Normal Pod simple-prod-query-d47cf7449-cl95w.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:17 +0000 UTC Normal Pod simple-prod-query-d47cf7449-cl95w.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:17 +0000 UTC Normal Pod simple-prod-query-d47cf7449-s24r8.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:17 +0000 UTC Normal Pod simple-prod-query-d47cf7449-s24r8.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:17 +0000 UTC Normal Pod simple-prod-query-d47cf7449-s24r8.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:17 +0000 UTC Normal ReplicaSet.apps simple-prod-query-d47cf7449 SuccessfulDelete Deleted pod: simple-prod-query-d47cf7449-s24r8 replicaset-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:17 +0000 UTC Normal ReplicaSet.apps simple-prod-query-d47cf7449 SuccessfulDelete Deleted pod: simple-prod-query-d47cf7449-cl95w replicaset-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:17 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-d47cf7449 to 0 from 2 deployment-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:18 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-lrbbm Binding Scheduled Successfully assigned kuttl-test-unique-humpback/simple-prod-query-797b8b4b84-lrbbm to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:18 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-lrbbm AddedInterface Add eth0 [10.131.0.31/23] from ovn-kubernetes logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:18 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-lrbbm.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:18 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-lrbbm.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:18 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-lrbbm.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:18 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-lrbbm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:18 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-lrbbm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:18 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-lrbbm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:18 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-lrbbm.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:18 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-v6nwg Binding Scheduled Successfully assigned kuttl-test-unique-humpback/simple-prod-query-797b8b4b84-v6nwg to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:18 +0000 UTC Normal ReplicaSet.apps simple-prod-query-797b8b4b84 SuccessfulCreate Created pod: simple-prod-query-797b8b4b84-lrbbm replicaset-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:18 +0000 UTC Normal ReplicaSet.apps simple-prod-query-797b8b4b84 SuccessfulCreate Created pod: simple-prod-query-797b8b4b84-v6nwg replicaset-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:18 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-797b8b4b84 to 2 deployment-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:19 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-lrbbm.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:19 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-lrbbm.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:19 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-v6nwg AddedInterface Add eth0 [10.128.2.42/23] from ovn-kubernetes logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:19 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-v6nwg.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:19 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-v6nwg.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:19 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-v6nwg.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:19 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-v6nwg.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:19 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-v6nwg.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:19 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-v6nwg.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:19 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-v6nwg.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:19 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-v6nwg.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:19 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-v6nwg.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:24 +0000 UTC Normal Pod check-span-m6jtp Binding Scheduled Successfully assigned kuttl-test-unique-humpback/check-span-m6jtp to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:24 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-m6jtp job-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:24 +0000 UTC Normal Pod report-span-2vsfv Binding Scheduled Successfully assigned kuttl-test-unique-humpback/report-span-2vsfv to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:24 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-2vsfv job-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:25 +0000 UTC Normal Pod check-span-m6jtp AddedInterface Add eth0 [10.131.0.32/23] from ovn-kubernetes logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:25 +0000 UTC Normal Pod check-span-m6jtp.spec.containers{asserts-container} Pulling Pulling image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:25 +0000 UTC Normal Pod report-span-2vsfv AddedInterface Add eth0 [10.128.2.43/23] from ovn-kubernetes logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:25 +0000 UTC Normal Pod report-span-2vsfv.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:25 +0000 UTC Normal Pod report-span-2vsfv.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:25 +0000 UTC Normal Pod report-span-2vsfv.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:27 +0000 UTC Normal Pod check-span-m6jtp.spec.containers{asserts-container} Pulled Successfully pulled image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" in 2.64s (2.64s including waiting) kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:27 +0000 UTC Normal Pod check-span-m6jtp.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:27 +0000 UTC Normal Pod check-span-m6jtp.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:30 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-2-6459cf58df SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-2-6459nm5lb replicaset-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-2-6459nm5lb Binding Scheduled Successfully assigned kuttl-test-unique-humpback/elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-2-6459nm5lb to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-2-6459nm5lb AddedInterface Add eth0 [10.128.2.44/23] from ovn-kubernetes logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-2-6459nm5lb.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-2-6459nm5lb.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-2-6459nm5lb.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-2-6459nm5lb.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-2-6459nm5lb.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-2-6459nm5lb.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-2 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestuniquehumpbacksimpleprod-2-6459cf58df to 1 deployment-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Pod simple-prod-collector-6fcf554844-7mb82.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Pod simple-prod-collector-6fcf554844-lgbnf.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-6fcf554844 SuccessfulDelete Deleted pod: simple-prod-collector-6fcf554844-7mb82 replicaset-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-6fcf554844 SuccessfulDelete Deleted pod: simple-prod-collector-6fcf554844-lgbnf replicaset-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled down replica set simple-prod-collector-6fcf554844 to 0 from 2 deployment-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Warning Endpoints simple-prod-collector FailedToUpdateEndpoint Failed to update endpoint kuttl-test-unique-humpback/simple-prod-collector: Operation cannot be fulfilled on endpoints "simple-prod-collector": the object has been modified; please apply your changes to the latest version and try again endpoint-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-lrbbm.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-lrbbm.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-lrbbm.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-v6nwg.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-v6nwg.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Pod simple-prod-query-797b8b4b84-v6nwg.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal ReplicaSet.apps simple-prod-query-797b8b4b84 SuccessfulDelete Deleted pod: simple-prod-query-797b8b4b84-v6nwg replicaset-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal ReplicaSet.apps simple-prod-query-797b8b4b84 SuccessfulDelete Deleted pod: simple-prod-query-797b8b4b84-lrbbm replicaset-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:32 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-797b8b4b84 to 0 from 2 deployment-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:33 +0000 UTC Normal Pod simple-prod-collector-66b796fcbc-dm4rz Binding Scheduled Successfully assigned kuttl-test-unique-humpback/simple-prod-collector-66b796fcbc-dm4rz to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:33 +0000 UTC Normal Pod simple-prod-collector-66b796fcbc-kf4vf Binding Scheduled Successfully assigned kuttl-test-unique-humpback/simple-prod-collector-66b796fcbc-kf4vf to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:33 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-66b796fcbc SuccessfulCreate Created pod: simple-prod-collector-66b796fcbc-kf4vf replicaset-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:33 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-66b796fcbc SuccessfulCreate Created pod: simple-prod-collector-66b796fcbc-dm4rz replicaset-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:33 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-66b796fcbc to 2 deployment-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:33 +0000 UTC Normal Pod simple-prod-query-596466f689-5vz52 Binding Scheduled Successfully assigned kuttl-test-unique-humpback/simple-prod-query-596466f689-5vz52 to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:33 +0000 UTC Normal Pod simple-prod-query-596466f689-5vz52 AddedInterface Add eth0 [10.131.0.33/23] from ovn-kubernetes logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:33 +0000 UTC Normal Pod simple-prod-query-596466f689-5vz52.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:33 +0000 UTC Normal Pod simple-prod-query-596466f689-nvvfx Binding Scheduled Successfully assigned kuttl-test-unique-humpback/simple-prod-query-596466f689-nvvfx to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:33 +0000 UTC Normal ReplicaSet.apps simple-prod-query-596466f689 SuccessfulCreate Created pod: simple-prod-query-596466f689-5vz52 replicaset-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:33 +0000 UTC Normal ReplicaSet.apps simple-prod-query-596466f689 SuccessfulCreate Created pod: simple-prod-query-596466f689-nvvfx replicaset-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:33 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-596466f689 to 2 deployment-controller logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-collector-66b796fcbc-dm4rz AddedInterface Add eth0 [10.129.2.21/23] from ovn-kubernetes logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-collector-66b796fcbc-dm4rz.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-collector-66b796fcbc-kf4vf AddedInterface Add eth0 [10.131.0.34/23] from ovn-kubernetes logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-collector-66b796fcbc-kf4vf.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-collector-66b796fcbc-kf4vf.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-collector-66b796fcbc-kf4vf.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-query-596466f689-5vz52.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-query-596466f689-5vz52.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-query-596466f689-5vz52.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-query-596466f689-5vz52.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-query-596466f689-5vz52.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-query-596466f689-5vz52.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-query-596466f689-5vz52.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-query-596466f689-5vz52.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-query-596466f689-nvvfx AddedInterface Add eth0 [10.128.2.45/23] from ovn-kubernetes logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-query-596466f689-nvvfx.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-query-596466f689-nvvfx.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-query-596466f689-nvvfx.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-query-596466f689-nvvfx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-query-596466f689-nvvfx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-query-596466f689-nvvfx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-query-596466f689-nvvfx.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-query-596466f689-nvvfx.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:34 +0000 UTC Normal Pod simple-prod-query-596466f689-nvvfx.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:05:36 | es-increasing-replicas | 2024-01-08 07:05:36 +0000 UTC Normal Pod simple-prod-collector-66b796fcbc-dm4rz.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" in 2.183s (2.183s including waiting) kubelet logger.go:42: 07:05:36 | es-increasing-replicas | Deleting namespace: kuttl-test-unique-humpback === CONT kuttl/harness/es-index-cleaner-autoprov logger.go:42: 07:06:08 | es-index-cleaner-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:06:08 | es-index-cleaner-autoprov | Creating namespace: kuttl-test-promoted-tuna logger.go:42: 07:06:08 | es-index-cleaner-autoprov/1-install | starting test step 1-install logger.go:42: 07:06:08 | es-index-cleaner-autoprov/1-install | Jaeger:kuttl-test-promoted-tuna/test-es-index-cleaner-with-prefix created logger.go:42: 07:06:45 | es-index-cleaner-autoprov/1-install | test step completed 1-install logger.go:42: 07:06:45 | es-index-cleaner-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:06:45 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE test-es-index-cleaner-with-prefix /dev/null] logger.go:42: 07:06:47 | es-index-cleaner-autoprov/2-report-spans | Warning: resource jaegers/test-es-index-cleaner-with-prefix is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:06:53 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c DAYS=5 ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:06:54 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:06:54 | es-index-cleaner-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:07:33 | es-index-cleaner-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 07:07:33 | es-index-cleaner-autoprov/3-install | starting test step 3-install logger.go:42: 07:07:33 | es-index-cleaner-autoprov/3-install | Jaeger:kuttl-test-promoted-tuna/test-es-index-cleaner-with-prefix updated logger.go:42: 07:07:33 | es-index-cleaner-autoprov/3-install | test step completed 3-install logger.go:42: 07:07:33 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | starting test step 4-wait-es-index-cleaner logger.go:42: 07:07:33 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob test-es-index-cleaner-with-prefix-es-index-cleaner --namespace $NAMESPACE] logger.go:42: 07:07:34 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-08T07:07:34Z" level=debug msg="Checking if the test-es-index-cleaner-with-prefix-es-index-cleaner CronJob exists" logger.go:42: 07:07:34 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-08T07:07:34Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 07:07:34 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-08T07:07:34Z" level=warning msg="The BatchV1/Cronjob test-es-index-cleaner-with-prefix-es-index-cleaner was not found" logger.go:42: 07:07:34 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-08T07:07:34Z" level=debug msg="Found BatchV/Cronjobs:" logger.go:42: 07:07:44 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-08T07:07:44Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 07:07:44 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-08T07:07:44Z" level=info msg="Cronjob test-es-index-cleaner-with-prefix-es-index-cleaner found successfully" logger.go:42: 07:07:44 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-08T07:07:44Z" level=debug msg="Waiting for the next scheduled job from test-es-index-cleaner-with-prefix-es-index-cleaner cronjob" logger.go:42: 07:07:44 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-08T07:07:44Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 07:07:54 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-08T07:07:54Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 07:08:04 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-08T07:08:04Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 07:08:14 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-08T07:08:14Z" level=info msg="Job of owner test-es-index-cleaner-with-prefix-es-index-cleaner succeeded after test-es-index-cleaner-with-prefix-es-index-cleaner 30.038316267s" logger.go:42: 07:08:14 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | test step completed 4-wait-es-index-cleaner logger.go:42: 07:08:14 | es-index-cleaner-autoprov/5-install | starting test step 5-install logger.go:42: 07:08:14 | es-index-cleaner-autoprov/5-install | Jaeger:kuttl-test-promoted-tuna/test-es-index-cleaner-with-prefix updated logger.go:42: 07:08:14 | es-index-cleaner-autoprov/5-install | test step completed 5-install logger.go:42: 07:08:14 | es-index-cleaner-autoprov/6-check-indices | starting test step 6-check-indices logger.go:42: 07:08:14 | es-index-cleaner-autoprov/6-check-indices | Job:kuttl-test-promoted-tuna/00-check-indices created logger.go:42: 07:08:18 | es-index-cleaner-autoprov/6-check-indices | test step completed 6-check-indices logger.go:42: 07:08:18 | es-index-cleaner-autoprov | es-index-cleaner-autoprov events from ns kuttl-test-promoted-tuna: logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:16 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestpromotedtunatestesindexclea-1-85646bd9cf SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestpromotedtunatestesindexclea-1-8c7gpl replicaset-controller logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromotedtunatestesindexclea-1-8c7gpl Binding Scheduled Successfully assigned kuttl-test-promoted-tuna/elasticsearch-cdm-kuttltestpromotedtunatestesindexclea-1-8c7gpl to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromotedtunatestesindexclea-1-8c7gpl AddedInterface Add eth0 [10.129.2.22/23] from ovn-kubernetes logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromotedtunatestesindexclea-1-8c7gpl.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromotedtunatestesindexclea-1-8c7gpl.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromotedtunatestesindexclea-1-8c7gpl.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromotedtunatestesindexclea-1-8c7gpl.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:16 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestpromotedtunatestesindexclea-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestpromotedtunatestesindexclea-1-85646bd9cf to 1 deployment-controller logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromotedtunatestesindexclea-1-8c7gpl.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromotedtunatestesindexclea-1-8c7gpl.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:26 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpromotedtunatestesindexclea-1-8c7gpl.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:31 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpromotedtunatestesindexclea-1-8c7gpl.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-848857dc86-9b7g2 Binding Scheduled Successfully assigned kuttl-test-promoted-tuna/test-es-index-cleaner-with-prefix-collector-848857dc86-9b7g2 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-848857dc86-9b7g2 AddedInterface Add eth0 [10.128.2.46/23] from ovn-kubernetes logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-848857dc86-9b7g2.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-848857dc86-9b7g2.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-848857dc86-9b7g2.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-collector-848857dc86 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-collector-848857dc86-9b7g2 replicaset-controller logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-collector ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-collector-848857dc86 to 1 deployment-controller logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-955f9f89f-g8zkc Binding Scheduled Successfully assigned kuttl-test-promoted-tuna/test-es-index-cleaner-with-prefix-query-955f9f89f-g8zkc to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-955f9f89f-g8zkc AddedInterface Add eth0 [10.131.0.35/23] from ovn-kubernetes logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-955f9f89f-g8zkc.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-955f9f89f-g8zkc.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-955f9f89f-g8zkc.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-955f9f89f-g8zkc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-955f9f89f-g8zkc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-955f9f89f-g8zkc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-955f9f89f-g8zkc.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-955f9f89f-g8zkc.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-955f9f89f-g8zkc.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-955f9f89f SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-955f9f89f-g8zkc replicaset-controller logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:43 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-955f9f89f to 1 deployment-controller logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:48 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-955f9f89f-g8zkc.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:48 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-955f9f89f-g8zkc.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:48 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-955f9f89f-g8zkc.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:48 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-955f9f89f SuccessfulDelete Deleted pod: test-es-index-cleaner-with-prefix-query-955f9f89f-g8zkc replicaset-controller logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:48 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled down replica set test-es-index-cleaner-with-prefix-query-955f9f89f to 0 from 1 deployment-controller logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7fd458fc78-fhfxt Binding Scheduled Successfully assigned kuttl-test-promoted-tuna/test-es-index-cleaner-with-prefix-query-7fd458fc78-fhfxt to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7fd458fc78-fhfxt AddedInterface Add eth0 [10.131.0.36/23] from ovn-kubernetes logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7fd458fc78-fhfxt.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:49 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-7fd458fc78 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-7fd458fc78-fhfxt replicaset-controller logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:49 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-7fd458fc78 to 1 deployment-controller logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:50 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7fd458fc78-fhfxt.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:50 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7fd458fc78-fhfxt.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:50 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7fd458fc78-fhfxt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:50 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7fd458fc78-fhfxt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:50 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7fd458fc78-fhfxt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:50 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7fd458fc78-fhfxt.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:50 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7fd458fc78-fhfxt.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:50 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7fd458fc78-fhfxt.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:54 +0000 UTC Normal Pod 00-report-span-bpxhw Binding Scheduled Successfully assigned kuttl-test-promoted-tuna/00-report-span-bpxhw to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:54 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-bpxhw job-controller logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:55 +0000 UTC Normal Pod 00-report-span-bpxhw AddedInterface Add eth0 [10.128.2.47/23] from ovn-kubernetes logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:55 +0000 UTC Normal Pod 00-report-span-bpxhw.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:55 +0000 UTC Normal Pod 00-report-span-bpxhw.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:55 +0000 UTC Normal Pod 00-report-span-bpxhw.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:06:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:07:28 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:07:28 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:07:32 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:07:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-848857dc86-9b7g2 horizontal-pod-autoscaler logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:08:00 +0000 UTC Normal Job.batch test-es-index-cleaner-with-prefix-es-index-cleaner-28411628 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-es-index-cleaner-2841162fj452 job-controller logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:08:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2841162fj452 Binding Scheduled Successfully assigned kuttl-test-promoted-tuna/test-es-index-cleaner-with-prefix-es-index-cleaner-2841162fj452 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:08:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2841162fj452 AddedInterface Add eth0 [10.128.2.48/23] from ovn-kubernetes logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:08:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2841162fj452.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:e5bea9a64ae418869cfb556d70e0a586c21589aed8606b4ff5850780ff5bbbd6" kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:08:00 +0000 UTC Normal CronJob.batch test-es-index-cleaner-with-prefix-es-index-cleaner SuccessfulCreate Created job test-es-index-cleaner-with-prefix-es-index-cleaner-28411628 cronjob-controller logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:08:03 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2841162fj452.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:e5bea9a64ae418869cfb556d70e0a586c21589aed8606b4ff5850780ff5bbbd6" in 2.436s (2.436s including waiting) kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:08:03 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2841162fj452.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Created Created container test-es-index-cleaner-with-prefix-es-index-cleaner kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:08:03 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2841162fj452.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Started Started container test-es-index-cleaner-with-prefix-es-index-cleaner kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:08:05 +0000 UTC Normal Job.batch test-es-index-cleaner-with-prefix-es-index-cleaner-28411628 Completed Job completed job-controller logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:08:05 +0000 UTC Normal CronJob.batch test-es-index-cleaner-with-prefix-es-index-cleaner SawCompletedJob Saw completed job: test-es-index-cleaner-with-prefix-es-index-cleaner-28411628, status: Complete cronjob-controller logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:08:14 +0000 UTC Normal Pod 00-check-indices-zmgb5 Binding Scheduled Successfully assigned kuttl-test-promoted-tuna/00-check-indices-zmgb5 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:08:14 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-zmgb5 job-controller logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:08:15 +0000 UTC Normal Pod 00-check-indices-zmgb5 AddedInterface Add eth0 [10.128.2.49/23] from ovn-kubernetes logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:08:15 +0000 UTC Normal Pod 00-check-indices-zmgb5.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:08:15 +0000 UTC Normal Pod 00-check-indices-zmgb5.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:08:15 +0000 UTC Normal Pod 00-check-indices-zmgb5.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:08:18 | es-index-cleaner-autoprov | 2024-01-08 07:08:17 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 07:08:18 | es-index-cleaner-autoprov | Deleting namespace: kuttl-test-promoted-tuna === CONT kuttl/harness/es-from-aio-to-production logger.go:42: 07:08:26 | es-from-aio-to-production | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:08:26 | es-from-aio-to-production | Creating namespace: kuttl-test-busy-shrimp logger.go:42: 07:08:26 | es-from-aio-to-production/0-install | starting test step 0-install logger.go:42: 07:08:26 | es-from-aio-to-production/0-install | Jaeger:kuttl-test-busy-shrimp/my-jaeger created logger.go:42: 07:08:35 | es-from-aio-to-production/0-install | test step completed 0-install logger.go:42: 07:08:35 | es-from-aio-to-production/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:08:35 | es-from-aio-to-production/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:08:37 | es-from-aio-to-production/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:08:44 | es-from-aio-to-production/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:08:44 | es-from-aio-to-production/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:08:45 | es-from-aio-to-production/1-smoke-test | job.batch/report-span created logger.go:42: 07:08:45 | es-from-aio-to-production/1-smoke-test | job.batch/check-span created logger.go:42: 07:08:57 | es-from-aio-to-production/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:08:57 | es-from-aio-to-production/3-install | starting test step 3-install logger.go:42: 07:08:57 | es-from-aio-to-production/3-install | Jaeger:kuttl-test-busy-shrimp/my-jaeger updated logger.go:42: 07:09:32 | es-from-aio-to-production/3-install | test step completed 3-install logger.go:42: 07:09:32 | es-from-aio-to-production/4-smoke-test | starting test step 4-smoke-test logger.go:42: 07:09:32 | es-from-aio-to-production/4-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:09:39 | es-from-aio-to-production/4-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:09:40 | es-from-aio-to-production/4-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:09:40 | es-from-aio-to-production/4-smoke-test | job.batch/report-span unchanged logger.go:42: 07:09:40 | es-from-aio-to-production/4-smoke-test | job.batch/check-span unchanged logger.go:42: 07:09:40 | es-from-aio-to-production/4-smoke-test | test step completed 4-smoke-test logger.go:42: 07:09:40 | es-from-aio-to-production | es-from-aio-to-production events from ns kuttl-test-busy-shrimp: logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:29 +0000 UTC Normal Pod my-jaeger-58d5b4d7cd-ql8qp Binding Scheduled Successfully assigned kuttl-test-busy-shrimp/my-jaeger-58d5b4d7cd-ql8qp to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:29 +0000 UTC Normal ReplicaSet.apps my-jaeger-58d5b4d7cd SuccessfulCreate Created pod: my-jaeger-58d5b4d7cd-ql8qp replicaset-controller logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:29 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-58d5b4d7cd to 1 deployment-controller logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:30 +0000 UTC Warning Pod my-jaeger-58d5b4d7cd-ql8qp FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-oauth-proxy-tls" : secret "my-jaeger-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:30 +0000 UTC Normal Pod my-jaeger-58d5b4d7cd-ql8qp AddedInterface Add eth0 [10.131.0.37/23] from ovn-kubernetes logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:31 +0000 UTC Normal Pod my-jaeger-58d5b4d7cd-ql8qp.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:32 +0000 UTC Normal Pod my-jaeger-58d5b4d7cd-ql8qp.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" in 1.931s (1.931s including waiting) kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:33 +0000 UTC Normal Pod my-jaeger-58d5b4d7cd-ql8qp.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:33 +0000 UTC Normal Pod my-jaeger-58d5b4d7cd-ql8qp.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:33 +0000 UTC Normal Pod my-jaeger-58d5b4d7cd-ql8qp.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:33 +0000 UTC Normal Pod my-jaeger-58d5b4d7cd-ql8qp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:33 +0000 UTC Normal Pod my-jaeger-58d5b4d7cd-ql8qp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:39 +0000 UTC Normal Pod my-jaeger-58d5b4d7cd-ql8qp.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:39 +0000 UTC Normal Pod my-jaeger-58d5b4d7cd-ql8qp.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:39 +0000 UTC Normal ReplicaSet.apps my-jaeger-58d5b4d7cd SuccessfulDelete Deleted pod: my-jaeger-58d5b4d7cd-ql8qp replicaset-controller logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:39 +0000 UTC Normal Pod my-jaeger-d65976fd4-wdfsb Binding Scheduled Successfully assigned kuttl-test-busy-shrimp/my-jaeger-d65976fd4-wdfsb to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:39 +0000 UTC Normal ReplicaSet.apps my-jaeger-d65976fd4 SuccessfulCreate Created pod: my-jaeger-d65976fd4-wdfsb replicaset-controller logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:39 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-58d5b4d7cd to 0 from 1 deployment-controller logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:39 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-d65976fd4 to 1 deployment-controller logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:40 +0000 UTC Normal Pod my-jaeger-d65976fd4-wdfsb AddedInterface Add eth0 [10.128.2.50/23] from ovn-kubernetes logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:40 +0000 UTC Normal Pod my-jaeger-d65976fd4-wdfsb.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:41 +0000 UTC Normal Pod my-jaeger-d65976fd4-wdfsb.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" in 1.53s (1.53s including waiting) kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:41 +0000 UTC Normal Pod my-jaeger-d65976fd4-wdfsb.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:41 +0000 UTC Normal Pod my-jaeger-d65976fd4-wdfsb.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:41 +0000 UTC Normal Pod my-jaeger-d65976fd4-wdfsb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:41 +0000 UTC Normal Pod my-jaeger-d65976fd4-wdfsb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:41 +0000 UTC Normal Pod my-jaeger-d65976fd4-wdfsb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:45 +0000 UTC Normal Pod check-span-5v5r8 Binding Scheduled Successfully assigned kuttl-test-busy-shrimp/check-span-5v5r8 to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:45 +0000 UTC Normal Pod check-span-5v5r8 AddedInterface Add eth0 [10.131.0.38/23] from ovn-kubernetes logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:45 +0000 UTC Normal Pod check-span-5v5r8.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:45 +0000 UTC Normal Pod check-span-5v5r8.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:45 +0000 UTC Normal Pod check-span-5v5r8.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:45 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-5v5r8 job-controller logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:45 +0000 UTC Normal Pod report-span-8bqhm Binding Scheduled Successfully assigned kuttl-test-busy-shrimp/report-span-8bqhm to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:45 +0000 UTC Normal Pod report-span-8bqhm AddedInterface Add eth0 [10.129.2.23/23] from ovn-kubernetes logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:45 +0000 UTC Normal Pod report-span-8bqhm.spec.containers{report-span} Pulling Pulling image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:45 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-8bqhm job-controller logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:48 +0000 UTC Normal Pod report-span-8bqhm.spec.containers{report-span} Pulled Successfully pulled image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" in 3.049s (3.049s including waiting) kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:48 +0000 UTC Normal Pod report-span-8bqhm.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:48 +0000 UTC Normal Pod report-span-8bqhm.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:08:56 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusyshrimpmyjaeger-1-f55db7bd9-kz2bz Binding Scheduled Successfully assigned kuttl-test-busy-shrimp/elasticsearch-cdm-kuttltestbusyshrimpmyjaeger-1-f55db7bd9-kz2bz to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusyshrimpmyjaeger-1-f55db7bd9-kz2bz AddedInterface Add eth0 [10.129.2.24/23] from ovn-kubernetes logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusyshrimpmyjaeger-1-f55db7bd9-kz2bz.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusyshrimpmyjaeger-1-f55db7bd9-kz2bz.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusyshrimpmyjaeger-1-f55db7bd9-kz2bz.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusyshrimpmyjaeger-1-f55db7bd9-kz2bz.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:02 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestbusyshrimpmyjaeger-1-f55db7bd9 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestbusyshrimpmyjaeger-1-f55db7bd9-kz2bz replicaset-controller logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:02 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestbusyshrimpmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestbusyshrimpmyjaeger-1-f55db7bd9 to 1 deployment-controller logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusyshrimpmyjaeger-1-f55db7bd9-kz2bz.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusyshrimpmyjaeger-1-f55db7bd9-kz2bz.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:12 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestbusyshrimpmyjaeger-1-f55db7bd9-kz2bz.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:18 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestbusyshrimpmyjaeger-1-f55db7bd9-kz2bz.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:21 +0000 UTC Normal Job.batch report-span Completed Job completed job-controller logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal Pod my-jaeger-collector-66ffffd77b-cl5lz Binding Scheduled Successfully assigned kuttl-test-busy-shrimp/my-jaeger-collector-66ffffd77b-cl5lz to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal Pod my-jaeger-collector-66ffffd77b-cl5lz AddedInterface Add eth0 [10.131.0.39/23] from ovn-kubernetes logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal Pod my-jaeger-collector-66ffffd77b-cl5lz.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal Pod my-jaeger-collector-66ffffd77b-cl5lz.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal Pod my-jaeger-collector-66ffffd77b-cl5lz.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-66ffffd77b SuccessfulCreate Created pod: my-jaeger-collector-66ffffd77b-cl5lz replicaset-controller logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-66ffffd77b to 1 deployment-controller logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal Pod my-jaeger-query-6cbb579786-2ltnk Binding Scheduled Successfully assigned kuttl-test-busy-shrimp/my-jaeger-query-6cbb579786-2ltnk to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal Pod my-jaeger-query-6cbb579786-2ltnk AddedInterface Add eth0 [10.131.0.40/23] from ovn-kubernetes logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal Pod my-jaeger-query-6cbb579786-2ltnk.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal Pod my-jaeger-query-6cbb579786-2ltnk.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal Pod my-jaeger-query-6cbb579786-2ltnk.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal Pod my-jaeger-query-6cbb579786-2ltnk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal Pod my-jaeger-query-6cbb579786-2ltnk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal Pod my-jaeger-query-6cbb579786-2ltnk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal Pod my-jaeger-query-6cbb579786-2ltnk.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal Pod my-jaeger-query-6cbb579786-2ltnk.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal Pod my-jaeger-query-6cbb579786-2ltnk.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-6cbb579786 SuccessfulCreate Created pod: my-jaeger-query-6cbb579786-2ltnk replicaset-controller logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:29 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-6cbb579786 to 1 deployment-controller logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:31 +0000 UTC Normal Pod my-jaeger-d65976fd4-wdfsb.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | 2024-01-08 07:09:31 +0000 UTC Normal Pod my-jaeger-d65976fd4-wdfsb.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:09:40 | es-from-aio-to-production | Deleting namespace: kuttl-test-busy-shrimp === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (686.21s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.24s) --- PASS: kuttl/harness/es-multiinstance (104.12s) --- PASS: kuttl/harness/es-simple-prod (6.23s) --- PASS: kuttl/harness/es-rollover-autoprov (242.57s) --- PASS: kuttl/harness/es-increasing-replicas (107.27s) --- PASS: kuttl/harness/es-index-cleaner-autoprov (137.93s) --- PASS: kuttl/harness/es-from-aio-to-production (81.69s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name elasticsearch --report --output /logs/artifacts/elasticsearch.xml ./artifacts/kuttl-report.xml time="2024-01-08T07:09:49Z" level=debug msg="Setting a new name for the test suites" time="2024-01-08T07:09:49Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-01-08T07:09:49Z" level=debug msg="normalizing test case names" time="2024-01-08T07:09:49Z" level=debug msg="elasticsearch/artifacts -> elasticsearch_artifacts" time="2024-01-08T07:09:49Z" level=debug msg="elasticsearch/es-multiinstance -> elasticsearch_es_multiinstance" time="2024-01-08T07:09:49Z" level=debug msg="elasticsearch/es-simple-prod -> elasticsearch_es_simple_prod" time="2024-01-08T07:09:49Z" level=debug msg="elasticsearch/es-rollover-autoprov -> elasticsearch_es_rollover_autoprov" time="2024-01-08T07:09:49Z" level=debug msg="elasticsearch/es-increasing-replicas -> elasticsearch_es_increasing_replicas" time="2024-01-08T07:09:49Z" level=debug msg="elasticsearch/es-index-cleaner-autoprov -> elasticsearch_es_index_cleaner_autoprov" time="2024-01-08T07:09:49Z" level=debug msg="elasticsearch/es-from-aio-to-production -> elasticsearch_es_from_aio_to_production" +-----------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------+--------+ | elasticsearch_artifacts | passed | | elasticsearch_es_multiinstance | passed | | elasticsearch_es_simple_prod | passed | | elasticsearch_es_rollover_autoprov | passed | | elasticsearch_es_increasing_replicas | passed | | elasticsearch_es_index_cleaner_autoprov | passed | | elasticsearch_es_from_aio_to_production | passed | +-----------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh examples false true + '[' 3 -ne 3 ']' + test_suite_name=examples + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/examples.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-examples make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ VERTX_IMG=jaegertracing/vertx-create-span:operator-e2e-tests \ ./tests/e2e/examples/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-06-062415 True False 18m Cluster version is 4.15.0-0.nightly-2024-01-06-062415' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-06-062415 True False 18m Cluster version is 4.15.0-0.nightly-2024-01-06-062415' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/examples/render.sh ++ export SUITE_DIR=./tests/e2e/examples ++ SUITE_DIR=./tests/e2e/examples ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/examples ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test examples-agent-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-agent-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-with-priority-class\e[0m' Rendering files for test examples-agent-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + mkdir -p examples-agent-with-priority-class + cd examples-agent-with-priority-class + example_name=agent-with-priority-class + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + '[' true '!=' true ']' + render_install_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + install_file=./02-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/agent-with-priority-class.yaml -o ./02-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./02-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./02-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./02-install.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + JAEGER_NAME=agent-as-daemonset + local jaeger_strategy ++ get_jaeger_strategy ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./02-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./02-install.yaml ++ strategy=DaemonSet ++ '[' DaemonSet = null ']' ++ echo DaemonSet ++ return 0 + jaeger_strategy=DaemonSet + '[' DaemonSet = DaemonSet ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./02-assert.yaml + render_smoke_test_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + jaeger_name=agent-as-daemonset + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test agent-as-daemonset true 02 + '[' 3 -ne 3 ']' + jaeger=agent-as-daemonset + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-all-in-one-with-options + '[' 1 -ne 1 ']' + test_name=examples-all-in-one-with-options + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-all-in-one-with-options' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-all-in-one-with-options\e[0m' Rendering files for test examples-all-in-one-with-options + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-with-priority-class + '[' examples-agent-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-all-in-one-with-options + cd examples-all-in-one-with-options + example_name=all-in-one-with-options + render_install_example all-in-one-with-options 00 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/all-in-one-with-options.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + JAEGER_NAME=my-jaeger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.metadata.name="my-jaeger"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i 'del(.spec.allInOne.image)' ./00-install.yaml + render_smoke_test_example all-in-one-with-options 01 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + jaeger_name=my-jaeger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + sed -i s~my-jaeger-query:443~my-jaeger-query:443/jaeger~gi ./01-smoke-test.yaml + start_test examples-business-application-injected-sidecar + '[' 1 -ne 1 ']' + test_name=examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-business-application-injected-sidecar' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-business-application-injected-sidecar\e[0m' Rendering files for test examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-all-in-one-with-options + '[' examples-all-in-one-with-options '!=' _build ']' + cd .. + mkdir -p examples-business-application-injected-sidecar + cd examples-business-application-injected-sidecar + example_name=simplest + cp /tmp/jaeger-tests/examples/business-application-injected-sidecar.yaml ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].image=strenv(VERTX_IMG)' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.path="/"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.port=8080' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.initialDelaySeconds=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.failureThreshold=3' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.periodSeconds=10' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.successThreshold=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.timeoutSeconds=1' ./00-install.yaml + render_install_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example simplest 02 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 02 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-collector-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-collector-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-collector-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-collector-with-priority-class\e[0m' Rendering files for test examples-collector-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-business-application-injected-sidecar + '[' examples-business-application-injected-sidecar '!=' _build ']' + cd .. + mkdir -p examples-collector-with-priority-class + cd examples-collector-with-priority-class + example_name=collector-with-priority-class + render_install_example collector-with-priority-class 00 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/collector-with-priority-class.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + JAEGER_NAME=collector-with-high-priority + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example collector-with-priority-class 01 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + jaeger_name=collector-with-high-priority + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test collector-with-high-priority true 01 + '[' 3 -ne 3 ']' + jaeger=collector-with-high-priority + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + export JAEGER_NAME=collector-with-high-priority + JAEGER_NAME=collector-with-high-priority + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-service-types + '[' 1 -ne 1 ']' + test_name=examples-service-types + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-service-types' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-service-types\e[0m' Rendering files for test examples-service-types + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-collector-with-priority-class + '[' examples-collector-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-service-types + cd examples-service-types + example_name=service-types + render_install_example service-types 00 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/service-types.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + JAEGER_NAME=service-types + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example service-types 01 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/service-types.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/service-types.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/service-types.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/service-types.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + jaeger_name=service-types + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test service-types true 01 + '[' 3 -ne 3 ']' + jaeger=service-types + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + export JAEGER_NAME=service-types + JAEGER_NAME=service-types + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod + '[' 1 -ne 1 ']' + test_name=examples-simple-prod + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod\e[0m' Rendering files for test examples-simple-prod + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-service-types + '[' examples-service-types '!=' _build ']' + cd .. + mkdir -p examples-simple-prod + cd examples-simple-prod + example_name=simple-prod + render_install_example simple-prod 01 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod 02 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod-with-volumes + '[' 1 -ne 1 ']' + test_name=examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod-with-volumes' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod-with-volumes\e[0m' Rendering files for test examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod + '[' examples-simple-prod '!=' _build ']' + cd .. + mkdir -p examples-simple-prod-with-volumes + cd examples-simple-prod-with-volumes + example_name=simple-prod-with-volumes + render_install_example simple-prod-with-volumes 01 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod-with-volumes 02 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + /tmp/jaeger-tests/bin/gomplate -f ./03-check-volume.yaml.template -o 03-check-volume.yaml + start_test examples-simplest + '[' 1 -ne 1 ']' + test_name=examples-simplest + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simplest' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simplest\e[0m' Rendering files for test examples-simplest + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod-with-volumes + '[' examples-simple-prod-with-volumes '!=' _build ']' + cd .. + mkdir -p examples-simplest + cd examples-simplest + example_name=simplest + render_install_example simplest 00 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 01 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger + '[' 1 -ne 1 ']' + test_name=examples-with-badger + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger\e[0m' Rendering files for test examples-with-badger + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simplest + '[' examples-simplest '!=' _build ']' + cd .. + mkdir -p examples-with-badger + cd examples-with-badger + example_name=with-badger + render_install_example with-badger 00 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + JAEGER_NAME=with-badger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger 01 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + jaeger_name=with-badger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + export JAEGER_NAME=with-badger + JAEGER_NAME=with-badger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger-and-volume + '[' 1 -ne 1 ']' + test_name=examples-with-badger-and-volume + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger-and-volume' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger-and-volume\e[0m' Rendering files for test examples-with-badger-and-volume + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger + '[' examples-with-badger '!=' _build ']' + cd .. + mkdir -p examples-with-badger-and-volume + cd examples-with-badger-and-volume + example_name=with-badger-and-volume + render_install_example with-badger-and-volume 00 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger-and-volume.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + JAEGER_NAME=with-badger-and-volume + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger-and-volume 01 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + jaeger_name=with-badger-and-volume + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger-and-volume true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger-and-volume + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + export JAEGER_NAME=with-badger-and-volume + JAEGER_NAME=with-badger-and-volume + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-cassandra + '[' 1 -ne 1 ']' + test_name=examples-with-cassandra + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-cassandra' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-cassandra\e[0m' Rendering files for test examples-with-cassandra + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger-and-volume + '[' examples-with-badger-and-volume '!=' _build ']' + cd .. + mkdir -p examples-with-cassandra + cd examples-with-cassandra + example_name=with-cassandra + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-cassandra 01 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-cassandra.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + JAEGER_NAME=with-cassandra + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-cassandra 02 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-cassandra.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-cassandra.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + jaeger_name=with-cassandra + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-cassandra true 02 + '[' 3 -ne 3 ']' + jaeger=with-cassandra + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + export JAEGER_NAME=with-cassandra + JAEGER_NAME=with-cassandra + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-sampling + '[' 1 -ne 1 ']' + test_name=examples-with-sampling + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-sampling' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-sampling\e[0m' Rendering files for test examples-with-sampling + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-cassandra + '[' examples-with-cassandra '!=' _build ']' + cd .. + mkdir -p examples-with-sampling + cd examples-with-sampling + export example_name=with-sampling + example_name=with-sampling + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-sampling 01 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-sampling.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + JAEGER_NAME=with-sampling + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-sampling 02 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-sampling.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-sampling.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + jaeger_name=with-sampling + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-sampling true 02 + '[' 3 -ne 3 ']' + jaeger=with-sampling + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + export JAEGER_NAME=with-sampling + JAEGER_NAME=with-sampling + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-agent-as-daemonset + '[' 1 -ne 1 ']' + test_name=examples-agent-as-daemonset + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-as-daemonset' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-as-daemonset\e[0m' Rendering files for test examples-agent-as-daemonset + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-sampling + '[' examples-with-sampling '!=' _build ']' + cd .. + mkdir -p examples-agent-as-daemonset + cd examples-agent-as-daemonset + '[' true = true ']' + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/agent-as-daemonset.yaml -o 02-install.yaml + '[' true = true ']' + start_test examples-openshift-with-htpasswd + '[' 1 -ne 1 ']' + test_name=examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-openshift-with-htpasswd' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-openshift-with-htpasswd\e[0m' Rendering files for test examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-as-daemonset + '[' examples-agent-as-daemonset '!=' _build ']' + cd .. + mkdir -p examples-openshift-with-htpasswd + cd examples-openshift-with-htpasswd + export JAEGER_NAME=with-htpasswd + JAEGER_NAME=with-htpasswd + export JAEGER_USERNAME=awesomeuser + JAEGER_USERNAME=awesomeuser + export JAEGER_PASSWORD=awesomepassword + JAEGER_PASSWORD=awesomepassword + export 'JAEGER_USER_PASSWORD_HASH=awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' + JAEGER_USER_PASSWORD_HASH='awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ echo 'awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ base64 + SECRET=YXdlc29tZXVzZXI6e1NIQX11VWRxUFZVeXFOQm1FUlUwUXhqM0tGYVpuanc9Cg== + /tmp/jaeger-tests/bin/gomplate -f ./00-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/with-htpasswd.yaml -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + export 'GET_URL_COMMAND=kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + export 'URL=https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + INSECURE=true + JAEGER_USERNAME= + JAEGER_PASSWORD= + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-unsecured.yaml + JAEGER_USERNAME=wronguser + JAEGER_PASSWORD=wrongpassword + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-check-unauthorized.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./04-check-authorized.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running examples E2E tests' Running examples E2E tests + cd tests/e2e/examples/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3521674980 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 15 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/examples-agent-as-daemonset === PAUSE kuttl/harness/examples-agent-as-daemonset === RUN kuttl/harness/examples-agent-with-priority-class === PAUSE kuttl/harness/examples-agent-with-priority-class === RUN kuttl/harness/examples-all-in-one-with-options === PAUSE kuttl/harness/examples-all-in-one-with-options === RUN kuttl/harness/examples-business-application-injected-sidecar === PAUSE kuttl/harness/examples-business-application-injected-sidecar === RUN kuttl/harness/examples-collector-with-priority-class === PAUSE kuttl/harness/examples-collector-with-priority-class === RUN kuttl/harness/examples-openshift-with-htpasswd === PAUSE kuttl/harness/examples-openshift-with-htpasswd === RUN kuttl/harness/examples-service-types === PAUSE kuttl/harness/examples-service-types === RUN kuttl/harness/examples-simple-prod === PAUSE kuttl/harness/examples-simple-prod === RUN kuttl/harness/examples-simple-prod-with-volumes === PAUSE kuttl/harness/examples-simple-prod-with-volumes === RUN kuttl/harness/examples-simplest === PAUSE kuttl/harness/examples-simplest === RUN kuttl/harness/examples-with-badger === PAUSE kuttl/harness/examples-with-badger === RUN kuttl/harness/examples-with-badger-and-volume === PAUSE kuttl/harness/examples-with-badger-and-volume === RUN kuttl/harness/examples-with-cassandra === PAUSE kuttl/harness/examples-with-cassandra === RUN kuttl/harness/examples-with-sampling === PAUSE kuttl/harness/examples-with-sampling === CONT kuttl/harness/artifacts logger.go:42: 07:10:20 | artifacts | Creating namespace: kuttl-test-innocent-ape logger.go:42: 07:10:20 | artifacts | artifacts events from ns kuttl-test-innocent-ape: logger.go:42: 07:10:20 | artifacts | Deleting namespace: kuttl-test-innocent-ape === CONT kuttl/harness/examples-simple-prod logger.go:42: 07:10:26 | examples-simple-prod | Creating namespace: kuttl-test-darling-dogfish logger.go:42: 07:10:26 | examples-simple-prod/1-install | starting test step 1-install logger.go:42: 07:10:27 | examples-simple-prod/1-install | Jaeger:kuttl-test-darling-dogfish/simple-prod created logger.go:42: 07:11:02 | examples-simple-prod/1-install | test step completed 1-install logger.go:42: 07:11:02 | examples-simple-prod/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:11:02 | examples-simple-prod/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:11:04 | examples-simple-prod/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:11:11 | examples-simple-prod/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:11:12 | examples-simple-prod/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:11:13 | examples-simple-prod/2-smoke-test | job.batch/report-span created logger.go:42: 07:11:13 | examples-simple-prod/2-smoke-test | job.batch/check-span created logger.go:42: 07:11:24 | examples-simple-prod/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:11:24 | examples-simple-prod | examples-simple-prod events from ns kuttl-test-darling-dogfish: logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:32 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestdarlingdogfishsimpleprod-1-7fcf88854c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestdarlingdogfishsimpleprod-1-7fcf8w9m9 replicaset-controller logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdarlingdogfishsimpleprod-1-7fcf8w9m9 Binding Scheduled Successfully assigned kuttl-test-darling-dogfish/elasticsearch-cdm-kuttltestdarlingdogfishsimpleprod-1-7fcf8w9m9 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:32 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestdarlingdogfishsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestdarlingdogfishsimpleprod-1-7fcf88854c to 1 deployment-controller logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdarlingdogfishsimpleprod-1-7fcf8w9m9 AddedInterface Add eth0 [10.129.2.25/23] from ovn-kubernetes logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdarlingdogfishsimpleprod-1-7fcf8w9m9.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdarlingdogfishsimpleprod-1-7fcf8w9m9.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdarlingdogfishsimpleprod-1-7fcf8w9m9.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdarlingdogfishsimpleprod-1-7fcf8w9m9.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdarlingdogfishsimpleprod-1-7fcf8w9m9.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdarlingdogfishsimpleprod-1-7fcf8w9m9.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:48 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestdarlingdogfishsimpleprod-1-7fcf8w9m9.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:59 +0000 UTC Normal Pod simple-prod-collector-9c949d48c-ggbbj Binding Scheduled Successfully assigned kuttl-test-darling-dogfish/simple-prod-collector-9c949d48c-ggbbj to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:59 +0000 UTC Normal Pod simple-prod-collector-9c949d48c-ggbbj AddedInterface Add eth0 [10.128.2.51/23] from ovn-kubernetes logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:59 +0000 UTC Normal Pod simple-prod-collector-9c949d48c-ggbbj.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:59 +0000 UTC Normal Pod simple-prod-collector-9c949d48c-ggbbj.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:59 +0000 UTC Normal Pod simple-prod-collector-9c949d48c-ggbbj.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:59 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-9c949d48c SuccessfulCreate Created pod: simple-prod-collector-9c949d48c-ggbbj replicaset-controller logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:59 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-9c949d48c to 1 deployment-controller logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:59 +0000 UTC Normal Pod simple-prod-query-b457458b5-gfbwr Binding Scheduled Successfully assigned kuttl-test-darling-dogfish/simple-prod-query-b457458b5-gfbwr to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:59 +0000 UTC Normal Pod simple-prod-query-b457458b5-gfbwr AddedInterface Add eth0 [10.131.0.41/23] from ovn-kubernetes logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:59 +0000 UTC Normal Pod simple-prod-query-b457458b5-gfbwr.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:59 +0000 UTC Normal ReplicaSet.apps simple-prod-query-b457458b5 SuccessfulCreate Created pod: simple-prod-query-b457458b5-gfbwr replicaset-controller logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:10:59 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-b457458b5 to 1 deployment-controller logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:00 +0000 UTC Normal Pod simple-prod-query-b457458b5-gfbwr.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:00 +0000 UTC Normal Pod simple-prod-query-b457458b5-gfbwr.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:00 +0000 UTC Normal Pod simple-prod-query-b457458b5-gfbwr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:00 +0000 UTC Normal Pod simple-prod-query-b457458b5-gfbwr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:00 +0000 UTC Normal Pod simple-prod-query-b457458b5-gfbwr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:00 +0000 UTC Normal Pod simple-prod-query-b457458b5-gfbwr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:00 +0000 UTC Normal Pod simple-prod-query-b457458b5-gfbwr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:00 +0000 UTC Normal Pod simple-prod-query-b457458b5-gfbwr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:05 +0000 UTC Normal Pod simple-prod-query-b457458b5-gfbwr.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:05 +0000 UTC Normal Pod simple-prod-query-b457458b5-gfbwr.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:05 +0000 UTC Normal Pod simple-prod-query-b457458b5-gfbwr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:05 +0000 UTC Normal ReplicaSet.apps simple-prod-query-b457458b5 SuccessfulDelete Deleted pod: simple-prod-query-b457458b5-gfbwr replicaset-controller logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:05 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-b457458b5 to 0 from 1 deployment-controller logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:07 +0000 UTC Normal Pod simple-prod-query-64f7d67c5c-5qmc2 Binding Scheduled Successfully assigned kuttl-test-darling-dogfish/simple-prod-query-64f7d67c5c-5qmc2 to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:07 +0000 UTC Normal Pod simple-prod-query-64f7d67c5c-5qmc2 AddedInterface Add eth0 [10.131.0.42/23] from ovn-kubernetes logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:07 +0000 UTC Normal Pod simple-prod-query-64f7d67c5c-5qmc2.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:07 +0000 UTC Normal ReplicaSet.apps simple-prod-query-64f7d67c5c SuccessfulCreate Created pod: simple-prod-query-64f7d67c5c-5qmc2 replicaset-controller logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:07 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-64f7d67c5c to 1 deployment-controller logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:08 +0000 UTC Normal Pod simple-prod-query-64f7d67c5c-5qmc2.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:08 +0000 UTC Normal Pod simple-prod-query-64f7d67c5c-5qmc2.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:08 +0000 UTC Normal Pod simple-prod-query-64f7d67c5c-5qmc2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:08 +0000 UTC Normal Pod simple-prod-query-64f7d67c5c-5qmc2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:08 +0000 UTC Normal Pod simple-prod-query-64f7d67c5c-5qmc2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:08 +0000 UTC Normal Pod simple-prod-query-64f7d67c5c-5qmc2.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:08 +0000 UTC Normal Pod simple-prod-query-64f7d67c5c-5qmc2.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:08 +0000 UTC Normal Pod simple-prod-query-64f7d67c5c-5qmc2.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:13 +0000 UTC Normal Pod check-span-rg5kn Binding Scheduled Successfully assigned kuttl-test-darling-dogfish/check-span-rg5kn to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:13 +0000 UTC Normal Pod check-span-rg5kn AddedInterface Add eth0 [10.128.2.53/23] from ovn-kubernetes logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:13 +0000 UTC Normal Pod check-span-rg5kn.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:13 +0000 UTC Normal Pod check-span-rg5kn.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:13 +0000 UTC Normal Pod check-span-rg5kn.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:13 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-rg5kn job-controller logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:13 +0000 UTC Normal Pod report-span-jhgjd Binding Scheduled Successfully assigned kuttl-test-darling-dogfish/report-span-jhgjd to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:13 +0000 UTC Normal Pod report-span-jhgjd AddedInterface Add eth0 [10.128.2.52/23] from ovn-kubernetes logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:13 +0000 UTC Normal Pod report-span-jhgjd.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:13 +0000 UTC Normal Pod report-span-jhgjd.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:13 +0000 UTC Normal Pod report-span-jhgjd.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:13 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-jhgjd job-controller logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:11:24 | examples-simple-prod | 2024-01-08 07:11:24 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:11:24 | examples-simple-prod | Deleting namespace: kuttl-test-darling-dogfish === CONT kuttl/harness/examples-with-sampling logger.go:42: 07:12:11 | examples-with-sampling | Creating namespace: kuttl-test-pro-zebra logger.go:42: 07:12:11 | examples-with-sampling/0-install | starting test step 0-install logger.go:42: 07:12:11 | examples-with-sampling/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:12:11 | examples-with-sampling/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:12:11 | examples-with-sampling/0-install | >>>> Creating namespace kuttl-test-pro-zebra logger.go:42: 07:12:11 | examples-with-sampling/0-install | kubectl create namespace kuttl-test-pro-zebra 2>&1 | grep -v "already exists" || true logger.go:42: 07:12:11 | examples-with-sampling/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-pro-zebra 2>&1 | grep -v "already exists" || true logger.go:42: 07:12:12 | examples-with-sampling/0-install | service/cassandra created logger.go:42: 07:12:12 | examples-with-sampling/0-install | statefulset.apps/cassandra created logger.go:42: 07:12:12 | examples-with-sampling/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:12:23 | examples-with-sampling/0-install | test step completed 0-install logger.go:42: 07:12:23 | examples-with-sampling/1-install | starting test step 1-install logger.go:42: 07:12:24 | examples-with-sampling/1-install | Jaeger:kuttl-test-pro-zebra/with-sampling created logger.go:42: 07:12:30 | examples-with-sampling/1-install | test step completed 1-install logger.go:42: 07:12:30 | examples-with-sampling/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:12:30 | examples-with-sampling/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-sampling /dev/null] logger.go:42: 07:12:32 | examples-with-sampling/2-smoke-test | Warning: resource jaegers/with-sampling is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:12:39 | examples-with-sampling/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:12:40 | examples-with-sampling/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:12:40 | examples-with-sampling/2-smoke-test | job.batch/report-span created logger.go:42: 07:12:40 | examples-with-sampling/2-smoke-test | job.batch/check-span created logger.go:42: 07:12:53 | examples-with-sampling/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:12:53 | examples-with-sampling/3- | starting test step 3- logger.go:42: 07:12:53 | examples-with-sampling/3- | test step completed 3- logger.go:42: 07:12:53 | examples-with-sampling | examples-with-sampling events from ns kuttl-test-pro-zebra: logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:12 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-pro-zebra/cassandra-0 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:12 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:13 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.129.2.26/23] from ovn-kubernetes logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:13 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:17 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 4.132s (4.132s including waiting) kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:17 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:17 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:17 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-pro-zebra/cassandra-1 to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:17 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:18 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.131.0.43/23] from ovn-kubernetes logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:18 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:23 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 4.121s (4.121s including waiting) kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:23 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:23 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:27 +0000 UTC Normal Pod with-sampling-c86756dbd-kvfr9 Binding Scheduled Successfully assigned kuttl-test-pro-zebra/with-sampling-c86756dbd-kvfr9 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:27 +0000 UTC Normal Pod with-sampling-c86756dbd-kvfr9 AddedInterface Add eth0 [10.128.2.54/23] from ovn-kubernetes logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:27 +0000 UTC Normal Pod with-sampling-c86756dbd-kvfr9.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:27 +0000 UTC Normal Pod with-sampling-c86756dbd-kvfr9.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:27 +0000 UTC Normal Pod with-sampling-c86756dbd-kvfr9.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:27 +0000 UTC Normal Pod with-sampling-c86756dbd-kvfr9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:27 +0000 UTC Normal ReplicaSet.apps with-sampling-c86756dbd SuccessfulCreate Created pod: with-sampling-c86756dbd-kvfr9 replicaset-controller logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:27 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-c86756dbd to 1 deployment-controller logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:28 +0000 UTC Normal Pod with-sampling-c86756dbd-kvfr9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:28 +0000 UTC Normal Pod with-sampling-c86756dbd-kvfr9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:34 +0000 UTC Normal Pod with-sampling-c86756dbd-kvfr9.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:34 +0000 UTC Normal Pod with-sampling-c86756dbd-kvfr9.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:34 +0000 UTC Normal ReplicaSet.apps with-sampling-c86756dbd SuccessfulDelete Deleted pod: with-sampling-c86756dbd-kvfr9 replicaset-controller logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:34 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled down replica set with-sampling-c86756dbd to 0 from 1 deployment-controller logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:35 +0000 UTC Normal Pod with-sampling-866dd778bc-cj2x2 Binding Scheduled Successfully assigned kuttl-test-pro-zebra/with-sampling-866dd778bc-cj2x2 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:35 +0000 UTC Normal Pod with-sampling-866dd778bc-cj2x2 AddedInterface Add eth0 [10.128.2.55/23] from ovn-kubernetes logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:35 +0000 UTC Normal Pod with-sampling-866dd778bc-cj2x2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:35 +0000 UTC Normal ReplicaSet.apps with-sampling-866dd778bc SuccessfulCreate Created pod: with-sampling-866dd778bc-cj2x2 replicaset-controller logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:35 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-866dd778bc to 1 deployment-controller logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:36 +0000 UTC Normal Pod with-sampling-866dd778bc-cj2x2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:36 +0000 UTC Normal Pod with-sampling-866dd778bc-cj2x2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:36 +0000 UTC Normal Pod with-sampling-866dd778bc-cj2x2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:36 +0000 UTC Normal Pod with-sampling-866dd778bc-cj2x2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:36 +0000 UTC Normal Pod with-sampling-866dd778bc-cj2x2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:40 +0000 UTC Normal Pod check-span-b489c Binding Scheduled Successfully assigned kuttl-test-pro-zebra/check-span-b489c to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:40 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-b489c job-controller logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:40 +0000 UTC Normal Pod report-span-sl78k Binding Scheduled Successfully assigned kuttl-test-pro-zebra/report-span-sl78k to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:40 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-sl78k job-controller logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:41 +0000 UTC Normal Pod check-span-b489c AddedInterface Add eth0 [10.131.0.44/23] from ovn-kubernetes logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:41 +0000 UTC Normal Pod check-span-b489c.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:41 +0000 UTC Normal Pod check-span-b489c.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:41 +0000 UTC Normal Pod check-span-b489c.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:41 +0000 UTC Normal Pod report-span-sl78k AddedInterface Add eth0 [10.129.2.27/23] from ovn-kubernetes logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:41 +0000 UTC Normal Pod report-span-sl78k.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:41 +0000 UTC Normal Pod report-span-sl78k.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:41 +0000 UTC Normal Pod report-span-sl78k.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:12:53 | examples-with-sampling | 2024-01-08 07:12:52 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:12:53 | examples-with-sampling | Deleting namespace: kuttl-test-pro-zebra === CONT kuttl/harness/examples-with-cassandra logger.go:42: 07:13:12 | examples-with-cassandra | Creating namespace: kuttl-test-integral-alien logger.go:42: 07:13:12 | examples-with-cassandra/0-install | starting test step 0-install logger.go:42: 07:13:12 | examples-with-cassandra/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:13:12 | examples-with-cassandra/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:13:12 | examples-with-cassandra/0-install | >>>> Creating namespace kuttl-test-integral-alien logger.go:42: 07:13:12 | examples-with-cassandra/0-install | kubectl create namespace kuttl-test-integral-alien 2>&1 | grep -v "already exists" || true logger.go:42: 07:13:12 | examples-with-cassandra/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-integral-alien 2>&1 | grep -v "already exists" || true logger.go:42: 07:13:13 | examples-with-cassandra/0-install | service/cassandra created logger.go:42: 07:13:13 | examples-with-cassandra/0-install | statefulset.apps/cassandra created logger.go:42: 07:13:13 | examples-with-cassandra/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:13:16 | examples-with-cassandra/0-install | test step completed 0-install logger.go:42: 07:13:16 | examples-with-cassandra/1-install | starting test step 1-install logger.go:42: 07:13:16 | examples-with-cassandra/1-install | Jaeger:kuttl-test-integral-alien/with-cassandra created logger.go:42: 07:13:35 | examples-with-cassandra/1-install | test step completed 1-install logger.go:42: 07:13:35 | examples-with-cassandra/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:13:35 | examples-with-cassandra/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-cassandra /dev/null] logger.go:42: 07:13:37 | examples-with-cassandra/2-smoke-test | Warning: resource jaegers/with-cassandra is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:13:44 | examples-with-cassandra/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:13:45 | examples-with-cassandra/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:13:45 | examples-with-cassandra/2-smoke-test | job.batch/report-span created logger.go:42: 07:13:45 | examples-with-cassandra/2-smoke-test | job.batch/check-span created logger.go:42: 07:13:58 | examples-with-cassandra/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:13:58 | examples-with-cassandra | examples-with-cassandra events from ns kuttl-test-integral-alien: logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:13 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-integral-alien/cassandra-0 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:13 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.129.2.28/23] from ovn-kubernetes logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:13 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:13 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:13 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:13 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:14 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-integral-alien/cassandra-1 to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:14 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.131.0.45/23] from ovn-kubernetes logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:14 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:14 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:14 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:14 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:19 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-znh2d Binding Scheduled Successfully assigned kuttl-test-integral-alien/with-cassandra-cassandra-schema-job-znh2d to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:19 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-znh2d AddedInterface Add eth0 [10.128.2.56/23] from ovn-kubernetes logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:19 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job SuccessfulCreate Created pod: with-cassandra-cassandra-schema-job-znh2d job-controller logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:20 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-znh2d.spec.containers{with-cassandra-cassandra-schema-job} Pulling Pulling image "jaegertracing/jaeger-cassandra-schema:1.51.0" kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:24 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-znh2d.spec.containers{with-cassandra-cassandra-schema-job} Pulled Successfully pulled image "jaegertracing/jaeger-cassandra-schema:1.51.0" in 4.733s (4.733s including waiting) kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:24 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-znh2d.spec.containers{with-cassandra-cassandra-schema-job} Created Created container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:24 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-znh2d.spec.containers{with-cassandra-cassandra-schema-job} Started Started container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:31 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job Completed Job completed job-controller logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:32 +0000 UTC Normal Pod with-cassandra-d9dcb88d8-j7q2x Binding Scheduled Successfully assigned kuttl-test-integral-alien/with-cassandra-d9dcb88d8-j7q2x to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:32 +0000 UTC Normal ReplicaSet.apps with-cassandra-d9dcb88d8 SuccessfulCreate Created pod: with-cassandra-d9dcb88d8-j7q2x replicaset-controller logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:32 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-d9dcb88d8 to 1 deployment-controller logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:33 +0000 UTC Normal Pod with-cassandra-d9dcb88d8-j7q2x AddedInterface Add eth0 [10.128.2.57/23] from ovn-kubernetes logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:33 +0000 UTC Normal Pod with-cassandra-d9dcb88d8-j7q2x.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:33 +0000 UTC Normal Pod with-cassandra-d9dcb88d8-j7q2x.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:33 +0000 UTC Normal Pod with-cassandra-d9dcb88d8-j7q2x.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:33 +0000 UTC Normal Pod with-cassandra-d9dcb88d8-j7q2x.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:33 +0000 UTC Normal Pod with-cassandra-d9dcb88d8-j7q2x.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:33 +0000 UTC Normal Pod with-cassandra-d9dcb88d8-j7q2x.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:33 +0000 UTC Warning Pod with-cassandra-d9dcb88d8-j7q2x FailedMount MountVolume.SetUp failed for volume "with-cassandra-service-ca" : configmap references non-existent config key: service-ca.crt kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:38 +0000 UTC Normal Pod with-cassandra-d9dcb88d8-j7q2x.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:38 +0000 UTC Normal Pod with-cassandra-d9dcb88d8-j7q2x.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:38 +0000 UTC Normal ReplicaSet.apps with-cassandra-d9dcb88d8 SuccessfulDelete Deleted pod: with-cassandra-d9dcb88d8-j7q2x replicaset-controller logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:38 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled down replica set with-cassandra-d9dcb88d8 to 0 from 1 deployment-controller logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:39 +0000 UTC Normal Pod with-cassandra-6dc6cf7f49-55956 Binding Scheduled Successfully assigned kuttl-test-integral-alien/with-cassandra-6dc6cf7f49-55956 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:39 +0000 UTC Normal ReplicaSet.apps with-cassandra-6dc6cf7f49 SuccessfulCreate Created pod: with-cassandra-6dc6cf7f49-55956 replicaset-controller logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:39 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-6dc6cf7f49 to 1 deployment-controller logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:40 +0000 UTC Normal Pod with-cassandra-6dc6cf7f49-55956 AddedInterface Add eth0 [10.128.2.58/23] from ovn-kubernetes logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:40 +0000 UTC Normal Pod with-cassandra-6dc6cf7f49-55956.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:40 +0000 UTC Normal Pod with-cassandra-6dc6cf7f49-55956.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:40 +0000 UTC Normal Pod with-cassandra-6dc6cf7f49-55956.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:40 +0000 UTC Normal Pod with-cassandra-6dc6cf7f49-55956.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:40 +0000 UTC Normal Pod with-cassandra-6dc6cf7f49-55956.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:40 +0000 UTC Normal Pod with-cassandra-6dc6cf7f49-55956.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:45 +0000 UTC Normal Pod check-span-zmxb9 Binding Scheduled Successfully assigned kuttl-test-integral-alien/check-span-zmxb9 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:45 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-zmxb9 job-controller logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:45 +0000 UTC Normal Pod report-span-n66wr Binding Scheduled Successfully assigned kuttl-test-integral-alien/report-span-n66wr to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:45 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-n66wr job-controller logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:46 +0000 UTC Normal Pod check-span-zmxb9 AddedInterface Add eth0 [10.129.2.30/23] from ovn-kubernetes logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:46 +0000 UTC Normal Pod check-span-zmxb9.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:46 +0000 UTC Normal Pod check-span-zmxb9.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:46 +0000 UTC Normal Pod check-span-zmxb9.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:46 +0000 UTC Normal Pod report-span-n66wr AddedInterface Add eth0 [10.129.2.29/23] from ovn-kubernetes logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:46 +0000 UTC Normal Pod report-span-n66wr.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:46 +0000 UTC Normal Pod report-span-n66wr.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:46 +0000 UTC Normal Pod report-span-n66wr.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:13:58 | examples-with-cassandra | 2024-01-08 07:13:57 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:13:58 | examples-with-cassandra | Deleting namespace: kuttl-test-integral-alien === CONT kuttl/harness/examples-with-badger-and-volume logger.go:42: 07:14:11 | examples-with-badger-and-volume | Creating namespace: kuttl-test-calm-lacewing logger.go:42: 07:14:11 | examples-with-badger-and-volume/0-install | starting test step 0-install logger.go:42: 07:14:11 | examples-with-badger-and-volume/0-install | Jaeger:kuttl-test-calm-lacewing/with-badger-and-volume created logger.go:42: 07:14:18 | examples-with-badger-and-volume/0-install | test step completed 0-install logger.go:42: 07:14:18 | examples-with-badger-and-volume/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:14:18 | examples-with-badger-and-volume/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger-and-volume /dev/null] logger.go:42: 07:14:20 | examples-with-badger-and-volume/1-smoke-test | Warning: resource jaegers/with-badger-and-volume is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:14:27 | examples-with-badger-and-volume/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:14:27 | examples-with-badger-and-volume/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:14:28 | examples-with-badger-and-volume/1-smoke-test | job.batch/report-span created logger.go:42: 07:14:28 | examples-with-badger-and-volume/1-smoke-test | job.batch/check-span created logger.go:42: 07:14:39 | examples-with-badger-and-volume/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:14:39 | examples-with-badger-and-volume | examples-with-badger-and-volume events from ns kuttl-test-calm-lacewing: logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:14 +0000 UTC Normal Pod with-badger-and-volume-6c6479f579-wqk84 Binding Scheduled Successfully assigned kuttl-test-calm-lacewing/with-badger-and-volume-6c6479f579-wqk84 to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:14 +0000 UTC Normal ReplicaSet.apps with-badger-and-volume-6c6479f579 SuccessfulCreate Created pod: with-badger-and-volume-6c6479f579-wqk84 replicaset-controller logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:14 +0000 UTC Normal Deployment.apps with-badger-and-volume ScalingReplicaSet Scaled up replica set with-badger-and-volume-6c6479f579 to 1 deployment-controller logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:15 +0000 UTC Warning Pod with-badger-and-volume-6c6479f579-wqk84 FailedMount MountVolume.SetUp failed for volume "with-badger-and-volume-collector-tls-config-volume" : secret "with-badger-and-volume-collector-headless-tls" not found kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:15 +0000 UTC Normal Pod with-badger-and-volume-6c6479f579-wqk84 AddedInterface Add eth0 [10.131.0.46/23] from ovn-kubernetes logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:15 +0000 UTC Normal Pod with-badger-and-volume-6c6479f579-wqk84.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:16 +0000 UTC Normal Pod with-badger-and-volume-6c6479f579-wqk84.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:16 +0000 UTC Normal Pod with-badger-and-volume-6c6479f579-wqk84.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:16 +0000 UTC Normal Pod with-badger-and-volume-6c6479f579-wqk84.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:16 +0000 UTC Normal Pod with-badger-and-volume-6c6479f579-wqk84.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:16 +0000 UTC Normal Pod with-badger-and-volume-6c6479f579-wqk84.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:22 +0000 UTC Normal Pod with-badger-and-volume-6c6479f579-wqk84.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:22 +0000 UTC Normal Pod with-badger-and-volume-6c6479f579-wqk84.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:22 +0000 UTC Normal ReplicaSet.apps with-badger-and-volume-6c6479f579 SuccessfulDelete Deleted pod: with-badger-and-volume-6c6479f579-wqk84 replicaset-controller logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:22 +0000 UTC Normal Pod with-badger-and-volume-7667bfcdd4-xp8pn Binding Scheduled Successfully assigned kuttl-test-calm-lacewing/with-badger-and-volume-7667bfcdd4-xp8pn to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:22 +0000 UTC Normal ReplicaSet.apps with-badger-and-volume-7667bfcdd4 SuccessfulCreate Created pod: with-badger-and-volume-7667bfcdd4-xp8pn replicaset-controller logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:22 +0000 UTC Normal Deployment.apps with-badger-and-volume ScalingReplicaSet Scaled down replica set with-badger-and-volume-6c6479f579 to 0 from 1 deployment-controller logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:22 +0000 UTC Normal Deployment.apps with-badger-and-volume ScalingReplicaSet Scaled up replica set with-badger-and-volume-7667bfcdd4 to 1 deployment-controller logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:23 +0000 UTC Normal Pod with-badger-and-volume-7667bfcdd4-xp8pn AddedInterface Add eth0 [10.128.2.59/23] from ovn-kubernetes logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:23 +0000 UTC Normal Pod with-badger-and-volume-7667bfcdd4-xp8pn.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:23 +0000 UTC Normal Pod with-badger-and-volume-7667bfcdd4-xp8pn.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:23 +0000 UTC Normal Pod with-badger-and-volume-7667bfcdd4-xp8pn.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:23 +0000 UTC Normal Pod with-badger-and-volume-7667bfcdd4-xp8pn.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:23 +0000 UTC Normal Pod with-badger-and-volume-7667bfcdd4-xp8pn.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:23 +0000 UTC Normal Pod with-badger-and-volume-7667bfcdd4-xp8pn.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:28 +0000 UTC Normal Pod check-span-hlkdl Binding Scheduled Successfully assigned kuttl-test-calm-lacewing/check-span-hlkdl to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:28 +0000 UTC Normal Pod check-span-hlkdl AddedInterface Add eth0 [10.129.2.32/23] from ovn-kubernetes logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:28 +0000 UTC Normal Pod check-span-hlkdl.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:28 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-hlkdl job-controller logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:28 +0000 UTC Normal Pod report-span-hfwxf Binding Scheduled Successfully assigned kuttl-test-calm-lacewing/report-span-hfwxf to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:28 +0000 UTC Normal Pod report-span-hfwxf AddedInterface Add eth0 [10.129.2.31/23] from ovn-kubernetes logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:28 +0000 UTC Normal Pod report-span-hfwxf.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:28 +0000 UTC Normal Pod report-span-hfwxf.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:28 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-hfwxf job-controller logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:29 +0000 UTC Normal Pod check-span-hlkdl.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:29 +0000 UTC Normal Pod check-span-hlkdl.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:29 +0000 UTC Normal Pod report-span-hfwxf.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:14:39 | examples-with-badger-and-volume | 2024-01-08 07:14:39 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:14:39 | examples-with-badger-and-volume | Deleting namespace: kuttl-test-calm-lacewing === CONT kuttl/harness/examples-with-badger logger.go:42: 07:14:52 | examples-with-badger | Creating namespace: kuttl-test-set-buzzard logger.go:42: 07:14:52 | examples-with-badger/0-install | starting test step 0-install logger.go:42: 07:14:52 | examples-with-badger/0-install | Jaeger:kuttl-test-set-buzzard/with-badger created logger.go:42: 07:14:58 | examples-with-badger/0-install | test step completed 0-install logger.go:42: 07:14:58 | examples-with-badger/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:14:58 | examples-with-badger/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger /dev/null] logger.go:42: 07:15:00 | examples-with-badger/1-smoke-test | Warning: resource jaegers/with-badger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:15:07 | examples-with-badger/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:15:08 | examples-with-badger/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:15:09 | examples-with-badger/1-smoke-test | job.batch/report-span created logger.go:42: 07:15:09 | examples-with-badger/1-smoke-test | job.batch/check-span created logger.go:42: 07:15:21 | examples-with-badger/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:15:21 | examples-with-badger | examples-with-badger events from ns kuttl-test-set-buzzard: logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:14:56 +0000 UTC Normal Pod with-badger-696c668594-qmggr Binding Scheduled Successfully assigned kuttl-test-set-buzzard/with-badger-696c668594-qmggr to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:14:56 +0000 UTC Normal Pod with-badger-696c668594-qmggr AddedInterface Add eth0 [10.128.2.60/23] from ovn-kubernetes logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:14:56 +0000 UTC Normal Pod with-badger-696c668594-qmggr.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:14:56 +0000 UTC Normal Pod with-badger-696c668594-qmggr.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:14:56 +0000 UTC Normal Pod with-badger-696c668594-qmggr.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:14:56 +0000 UTC Normal Pod with-badger-696c668594-qmggr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:14:56 +0000 UTC Normal Pod with-badger-696c668594-qmggr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:14:56 +0000 UTC Normal Pod with-badger-696c668594-qmggr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:14:56 +0000 UTC Normal ReplicaSet.apps with-badger-696c668594 SuccessfulCreate Created pod: with-badger-696c668594-qmggr replicaset-controller logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:14:56 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-696c668594 to 1 deployment-controller logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:02 +0000 UTC Normal Pod with-badger-696c668594-qmggr.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:02 +0000 UTC Normal Pod with-badger-696c668594-qmggr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:02 +0000 UTC Normal ReplicaSet.apps with-badger-696c668594 SuccessfulDelete Deleted pod: with-badger-696c668594-qmggr replicaset-controller logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:02 +0000 UTC Normal Pod with-badger-7d88d495d6-pjx5f Binding Scheduled Successfully assigned kuttl-test-set-buzzard/with-badger-7d88d495d6-pjx5f to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:02 +0000 UTC Normal ReplicaSet.apps with-badger-7d88d495d6 SuccessfulCreate Created pod: with-badger-7d88d495d6-pjx5f replicaset-controller logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:02 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled down replica set with-badger-696c668594 to 0 from 1 deployment-controller logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:02 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-7d88d495d6 to 1 deployment-controller logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:03 +0000 UTC Normal Pod with-badger-7d88d495d6-pjx5f AddedInterface Add eth0 [10.131.0.47/23] from ovn-kubernetes logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:03 +0000 UTC Normal Pod with-badger-7d88d495d6-pjx5f.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:03 +0000 UTC Normal Pod with-badger-7d88d495d6-pjx5f.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:03 +0000 UTC Normal Pod with-badger-7d88d495d6-pjx5f.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:03 +0000 UTC Normal Pod with-badger-7d88d495d6-pjx5f.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:03 +0000 UTC Normal Pod with-badger-7d88d495d6-pjx5f.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:03 +0000 UTC Normal Pod with-badger-7d88d495d6-pjx5f.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:09 +0000 UTC Normal Pod check-span-dcm5t Binding Scheduled Successfully assigned kuttl-test-set-buzzard/check-span-dcm5t to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:09 +0000 UTC Normal Pod check-span-dcm5t AddedInterface Add eth0 [10.128.2.61/23] from ovn-kubernetes logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:09 +0000 UTC Normal Pod check-span-dcm5t.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:09 +0000 UTC Normal Pod check-span-dcm5t.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:09 +0000 UTC Normal Pod check-span-dcm5t.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:09 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-dcm5t job-controller logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:09 +0000 UTC Normal Pod report-span-ptnvv Binding Scheduled Successfully assigned kuttl-test-set-buzzard/report-span-ptnvv to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:09 +0000 UTC Normal Pod report-span-ptnvv AddedInterface Add eth0 [10.129.2.34/23] from ovn-kubernetes logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:09 +0000 UTC Normal Pod report-span-ptnvv.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:09 +0000 UTC Normal Pod report-span-ptnvv.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:09 +0000 UTC Normal Pod report-span-ptnvv.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:09 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-ptnvv job-controller logger.go:42: 07:15:21 | examples-with-badger | 2024-01-08 07:15:21 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:15:21 | examples-with-badger | Deleting namespace: kuttl-test-set-buzzard === CONT kuttl/harness/examples-simplest logger.go:42: 07:15:34 | examples-simplest | Creating namespace: kuttl-test-glad-cowbird logger.go:42: 07:15:34 | examples-simplest/0-install | starting test step 0-install logger.go:42: 07:15:34 | examples-simplest/0-install | Jaeger:kuttl-test-glad-cowbird/simplest created logger.go:42: 07:15:40 | examples-simplest/0-install | test step completed 0-install logger.go:42: 07:15:40 | examples-simplest/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:15:40 | examples-simplest/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:15:42 | examples-simplest/1-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:15:50 | examples-simplest/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:15:50 | examples-simplest/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:15:51 | examples-simplest/1-smoke-test | job.batch/report-span created logger.go:42: 07:15:51 | examples-simplest/1-smoke-test | job.batch/check-span created logger.go:42: 07:16:02 | examples-simplest/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:16:02 | examples-simplest | examples-simplest events from ns kuttl-test-glad-cowbird: logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:37 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-6b8f49cf4b to 1 deployment-controller logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:38 +0000 UTC Normal Pod simplest-6b8f49cf4b-xljbn Binding Scheduled Successfully assigned kuttl-test-glad-cowbird/simplest-6b8f49cf4b-xljbn to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:38 +0000 UTC Normal Pod simplest-6b8f49cf4b-xljbn AddedInterface Add eth0 [10.128.2.62/23] from ovn-kubernetes logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:38 +0000 UTC Normal Pod simplest-6b8f49cf4b-xljbn.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:38 +0000 UTC Normal Pod simplest-6b8f49cf4b-xljbn.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:38 +0000 UTC Normal Pod simplest-6b8f49cf4b-xljbn.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:38 +0000 UTC Normal Pod simplest-6b8f49cf4b-xljbn.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:38 +0000 UTC Normal Pod simplest-6b8f49cf4b-xljbn.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:38 +0000 UTC Normal Pod simplest-6b8f49cf4b-xljbn.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:38 +0000 UTC Normal ReplicaSet.apps simplest-6b8f49cf4b SuccessfulCreate Created pod: simplest-6b8f49cf4b-xljbn replicaset-controller logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:44 +0000 UTC Normal Pod simplest-6b8f49cf4b-xljbn.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:44 +0000 UTC Normal Pod simplest-6b8f49cf4b-xljbn.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:44 +0000 UTC Normal ReplicaSet.apps simplest-6b8f49cf4b SuccessfulDelete Deleted pod: simplest-6b8f49cf4b-xljbn replicaset-controller logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:44 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-6b8f49cf4b to 0 from 1 deployment-controller logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:45 +0000 UTC Normal Pod simplest-7dc6844c4-nkl7r Binding Scheduled Successfully assigned kuttl-test-glad-cowbird/simplest-7dc6844c4-nkl7r to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:45 +0000 UTC Normal Pod simplest-7dc6844c4-nkl7r AddedInterface Add eth0 [10.131.0.48/23] from ovn-kubernetes logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:45 +0000 UTC Normal Pod simplest-7dc6844c4-nkl7r.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:45 +0000 UTC Normal Pod simplest-7dc6844c4-nkl7r.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:45 +0000 UTC Normal Pod simplest-7dc6844c4-nkl7r.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:45 +0000 UTC Normal Pod simplest-7dc6844c4-nkl7r.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:45 +0000 UTC Normal Pod simplest-7dc6844c4-nkl7r.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:45 +0000 UTC Normal Pod simplest-7dc6844c4-nkl7r.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:45 +0000 UTC Normal ReplicaSet.apps simplest-7dc6844c4 SuccessfulCreate Created pod: simplest-7dc6844c4-nkl7r replicaset-controller logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:45 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-7dc6844c4 to 1 deployment-controller logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:51 +0000 UTC Normal Pod check-span-w9gnt Binding Scheduled Successfully assigned kuttl-test-glad-cowbird/check-span-w9gnt to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:51 +0000 UTC Normal Pod check-span-w9gnt AddedInterface Add eth0 [10.128.2.63/23] from ovn-kubernetes logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:51 +0000 UTC Normal Pod check-span-w9gnt.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:51 +0000 UTC Normal Pod check-span-w9gnt.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:51 +0000 UTC Normal Pod check-span-w9gnt.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:51 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-w9gnt job-controller logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:51 +0000 UTC Normal Pod report-span-hlvs7 Binding Scheduled Successfully assigned kuttl-test-glad-cowbird/report-span-hlvs7 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:51 +0000 UTC Normal Pod report-span-hlvs7 AddedInterface Add eth0 [10.129.2.35/23] from ovn-kubernetes logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:51 +0000 UTC Normal Pod report-span-hlvs7.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:51 +0000 UTC Normal Pod report-span-hlvs7.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:51 +0000 UTC Normal Pod report-span-hlvs7.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:15:51 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-hlvs7 job-controller logger.go:42: 07:16:02 | examples-simplest | 2024-01-08 07:16:02 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:16:02 | examples-simplest | Deleting namespace: kuttl-test-glad-cowbird === CONT kuttl/harness/examples-simple-prod-with-volumes logger.go:42: 07:16:15 | examples-simple-prod-with-volumes | Ignoring 03-check-volume.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:16:15 | examples-simple-prod-with-volumes | Creating namespace: kuttl-test-up-snake logger.go:42: 07:16:15 | examples-simple-prod-with-volumes/1-install | starting test step 1-install logger.go:42: 07:16:15 | examples-simple-prod-with-volumes/1-install | Jaeger:kuttl-test-up-snake/simple-prod created logger.go:42: 07:16:52 | examples-simple-prod-with-volumes/1-install | test step completed 1-install logger.go:42: 07:16:52 | examples-simple-prod-with-volumes/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:16:52 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:16:55 | examples-simple-prod-with-volumes/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:17:02 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:17:03 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:17:03 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/report-span created logger.go:42: 07:17:03 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/check-span created logger.go:42: 07:17:16 | examples-simple-prod-with-volumes/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:17:16 | examples-simple-prod-with-volumes/3-check-volume | starting test step 3-check-volume logger.go:42: 07:17:16 | examples-simple-prod-with-volumes/3-check-volume | running command: [sh -c kubectl exec $(kubectl get pods -n $NAMESPACE -l app=jaeger -l app.kubernetes.io/component=collector -o yaml | /tmp/jaeger-tests/bin/yq e '.items[0].metadata.name') -n $NAMESPACE -- ls /usr/share/elasticsearch/data] logger.go:42: 07:17:16 | examples-simple-prod-with-volumes/3-check-volume | test step completed 3-check-volume logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | examples-simple-prod-with-volumes events from ns kuttl-test-up-snake: logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:21 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupsnakesimpleprod-1-b6dd986b7-b498z Binding Scheduled Successfully assigned kuttl-test-up-snake/elasticsearch-cdm-kuttltestupsnakesimpleprod-1-b6dd986b7-b498z to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:21 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestupsnakesimpleprod-1-b6dd986b7-b498z FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:21 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestupsnakesimpleprod-1-b6dd986b7 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestupsnakesimpleprod-1-b6dd986b7-b498z replicaset-controller logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:21 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestupsnakesimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestupsnakesimpleprod-1-b6dd986b7 to 1 deployment-controller logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupsnakesimpleprod-1-b6dd986b7-b498z AddedInterface Add eth0 [10.129.2.36/23] from ovn-kubernetes logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupsnakesimpleprod-1-b6dd986b7-b498z.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupsnakesimpleprod-1-b6dd986b7-b498z.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupsnakesimpleprod-1-b6dd986b7-b498z.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupsnakesimpleprod-1-b6dd986b7-b498z.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupsnakesimpleprod-1-b6dd986b7-b498z.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupsnakesimpleprod-1-b6dd986b7-b498z.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:32 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestupsnakesimpleprod-1-b6dd986b7-b498z.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:37 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestupsnakesimpleprod-1-b6dd986b7-b498z.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:49 +0000 UTC Normal Pod simple-prod-collector-58857866d7-lf8v2 Binding Scheduled Successfully assigned kuttl-test-up-snake/simple-prod-collector-58857866d7-lf8v2 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:49 +0000 UTC Normal Pod simple-prod-collector-58857866d7-lf8v2 AddedInterface Add eth0 [10.128.2.64/23] from ovn-kubernetes logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:49 +0000 UTC Normal Pod simple-prod-collector-58857866d7-lf8v2.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:49 +0000 UTC Normal Pod simple-prod-collector-58857866d7-lf8v2.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:49 +0000 UTC Normal Pod simple-prod-collector-58857866d7-lf8v2.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:49 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-58857866d7 SuccessfulCreate Created pod: simple-prod-collector-58857866d7-lf8v2 replicaset-controller logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:49 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-58857866d7 to 1 deployment-controller logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:49 +0000 UTC Normal Pod simple-prod-query-78fdc58d47-bkdfh Binding Scheduled Successfully assigned kuttl-test-up-snake/simple-prod-query-78fdc58d47-bkdfh to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:49 +0000 UTC Normal Pod simple-prod-query-78fdc58d47-bkdfh AddedInterface Add eth0 [10.131.0.49/23] from ovn-kubernetes logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:49 +0000 UTC Normal Pod simple-prod-query-78fdc58d47-bkdfh.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:49 +0000 UTC Normal Pod simple-prod-query-78fdc58d47-bkdfh.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:49 +0000 UTC Normal Pod simple-prod-query-78fdc58d47-bkdfh.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:49 +0000 UTC Normal Pod simple-prod-query-78fdc58d47-bkdfh.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:49 +0000 UTC Normal ReplicaSet.apps simple-prod-query-78fdc58d47 SuccessfulCreate Created pod: simple-prod-query-78fdc58d47-bkdfh replicaset-controller logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:49 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-78fdc58d47 to 1 deployment-controller logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:50 +0000 UTC Normal Pod simple-prod-query-78fdc58d47-bkdfh.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:50 +0000 UTC Normal Pod simple-prod-query-78fdc58d47-bkdfh.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:50 +0000 UTC Normal Pod simple-prod-query-78fdc58d47-bkdfh.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:50 +0000 UTC Normal Pod simple-prod-query-78fdc58d47-bkdfh.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:50 +0000 UTC Normal Pod simple-prod-query-78fdc58d47-bkdfh.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:58 +0000 UTC Normal Pod simple-prod-query-78fdc58d47-bkdfh.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:58 +0000 UTC Normal Pod simple-prod-query-78fdc58d47-bkdfh.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:58 +0000 UTC Normal Pod simple-prod-query-78fdc58d47-bkdfh.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:58 +0000 UTC Normal ReplicaSet.apps simple-prod-query-78fdc58d47 SuccessfulDelete Deleted pod: simple-prod-query-78fdc58d47-bkdfh replicaset-controller logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:58 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-78fdc58d47 to 0 from 1 deployment-controller logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:59 +0000 UTC Normal Pod simple-prod-query-65b5fd87f7-4zsgr Binding Scheduled Successfully assigned kuttl-test-up-snake/simple-prod-query-65b5fd87f7-4zsgr to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:59 +0000 UTC Normal Pod simple-prod-query-65b5fd87f7-4zsgr AddedInterface Add eth0 [10.131.0.50/23] from ovn-kubernetes logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:59 +0000 UTC Normal Pod simple-prod-query-65b5fd87f7-4zsgr.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:59 +0000 UTC Normal Pod simple-prod-query-65b5fd87f7-4zsgr.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:59 +0000 UTC Normal Pod simple-prod-query-65b5fd87f7-4zsgr.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:59 +0000 UTC Normal Pod simple-prod-query-65b5fd87f7-4zsgr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:59 +0000 UTC Normal Pod simple-prod-query-65b5fd87f7-4zsgr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:59 +0000 UTC Normal Pod simple-prod-query-65b5fd87f7-4zsgr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:59 +0000 UTC Normal Pod simple-prod-query-65b5fd87f7-4zsgr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:59 +0000 UTC Normal Pod simple-prod-query-65b5fd87f7-4zsgr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:59 +0000 UTC Normal Pod simple-prod-query-65b5fd87f7-4zsgr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:59 +0000 UTC Normal ReplicaSet.apps simple-prod-query-65b5fd87f7 SuccessfulCreate Created pod: simple-prod-query-65b5fd87f7-4zsgr replicaset-controller logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:16:59 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-65b5fd87f7 to 1 deployment-controller logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:17:03 +0000 UTC Normal Pod check-span-wx55b Binding Scheduled Successfully assigned kuttl-test-up-snake/check-span-wx55b to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:17:03 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-wx55b job-controller logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:17:03 +0000 UTC Normal Pod report-span-wx9nh Binding Scheduled Successfully assigned kuttl-test-up-snake/report-span-wx9nh to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:17:03 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-wx9nh job-controller logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:17:04 +0000 UTC Normal Pod check-span-wx55b AddedInterface Add eth0 [10.128.2.66/23] from ovn-kubernetes logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:17:04 +0000 UTC Normal Pod check-span-wx55b.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:17:04 +0000 UTC Normal Pod check-span-wx55b.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:17:04 +0000 UTC Normal Pod check-span-wx55b.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:17:04 +0000 UTC Normal Pod report-span-wx9nh AddedInterface Add eth0 [10.128.2.65/23] from ovn-kubernetes logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:17:04 +0000 UTC Normal Pod report-span-wx9nh.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:17:04 +0000 UTC Normal Pod report-span-wx9nh.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:17:04 +0000 UTC Normal Pod report-span-wx9nh.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:17:07 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:17:07 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:17:07 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | 2024-01-08 07:17:15 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:17:16 | examples-simple-prod-with-volumes | Deleting namespace: kuttl-test-up-snake === CONT kuttl/harness/examples-openshift-with-htpasswd logger.go:42: 07:17:30 | examples-openshift-with-htpasswd | Ignoring 00-install.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:17:30 | examples-openshift-with-htpasswd | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:17:30 | examples-openshift-with-htpasswd | Creating namespace: kuttl-test-helped-vulture logger.go:42: 07:17:30 | examples-openshift-with-htpasswd/0-install | starting test step 0-install logger.go:42: 07:17:30 | examples-openshift-with-htpasswd/0-install | Secret:kuttl-test-helped-vulture/htpasswd created logger.go:42: 07:17:30 | examples-openshift-with-htpasswd/0-install | test step completed 0-install logger.go:42: 07:17:30 | examples-openshift-with-htpasswd/1-install | starting test step 1-install logger.go:42: 07:17:30 | examples-openshift-with-htpasswd/1-install | Jaeger:kuttl-test-helped-vulture/with-htpasswd created logger.go:42: 07:17:36 | examples-openshift-with-htpasswd/1-install | test step completed 1-install logger.go:42: 07:17:36 | examples-openshift-with-htpasswd/2-check-unsecured | starting test step 2-check-unsecured logger.go:42: 07:17:36 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [./ensure-ingress-host.sh] logger.go:42: 07:17:36 | examples-openshift-with-htpasswd/2-check-unsecured | Checking the Ingress host value was populated logger.go:42: 07:17:36 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 0 logger.go:42: 07:17:36 | examples-openshift-with-htpasswd/2-check-unsecured | Hostname is with-htpasswd-kuttl-test-helped-vulture.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:17:36 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:17:36 | examples-openshift-with-htpasswd/2-check-unsecured | Checking an expected HTTP response logger.go:42: 07:17:36 | examples-openshift-with-htpasswd/2-check-unsecured | Running in OpenShift logger.go:42: 07:17:36 | examples-openshift-with-htpasswd/2-check-unsecured | Not using any secret logger.go:42: 07:17:36 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 1/30 the https://with-htpasswd-kuttl-test-helped-vulture.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:17:36 | examples-openshift-with-htpasswd/2-check-unsecured | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:17:36 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 2/30 the https://with-htpasswd-kuttl-test-helped-vulture.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:17:36 | examples-openshift-with-htpasswd/2-check-unsecured | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 07:17:46 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 3/30 the https://with-htpasswd-kuttl-test-helped-vulture.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:17:46 | examples-openshift-with-htpasswd/2-check-unsecured | curl response asserted properly logger.go:42: 07:17:46 | examples-openshift-with-htpasswd/2-check-unsecured | test step completed 2-check-unsecured logger.go:42: 07:17:46 | examples-openshift-with-htpasswd/3-check-unauthorized | starting test step 3-check-unauthorized logger.go:42: 07:17:46 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:17:46 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking the Ingress host value was populated logger.go:42: 07:17:46 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 0 logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/3-check-unauthorized | Hostname is with-htpasswd-kuttl-test-helped-vulture.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [sh -c JAEGER_USERNAME=wronguser JAEGER_PASSWORD=wrongpassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking an expected HTTP response logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/3-check-unauthorized | Running in OpenShift logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/3-check-unauthorized | Using Jaeger basic authentication logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 1/30 the https://with-htpasswd-kuttl-test-helped-vulture.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/3-check-unauthorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 2/30 the https://with-htpasswd-kuttl-test-helped-vulture.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/3-check-unauthorized | curl response asserted properly logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/3-check-unauthorized | test step completed 3-check-unauthorized logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/4-check-authorized | starting test step 4-check-authorized logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/4-check-authorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/4-check-authorized | Checking the Ingress host value was populated logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/4-check-authorized | Try number 0 logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/4-check-authorized | Hostname is with-htpasswd-kuttl-test-helped-vulture.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/4-check-authorized | running command: [sh -c JAEGER_USERNAME=awesomeuser JAEGER_PASSWORD=awesomepassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE with-htpasswd] logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/4-check-authorized | Checking an expected HTTP response logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/4-check-authorized | Running in OpenShift logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/4-check-authorized | Using Jaeger basic authentication logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/4-check-authorized | Try number 1/30 the https://with-htpasswd-kuttl-test-helped-vulture.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/4-check-authorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/4-check-authorized | Try number 2/30 the https://with-htpasswd-kuttl-test-helped-vulture.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/4-check-authorized | curl response asserted properly logger.go:42: 07:17:47 | examples-openshift-with-htpasswd/4-check-authorized | test step completed 4-check-authorized logger.go:42: 07:17:47 | examples-openshift-with-htpasswd | examples-openshift-with-htpasswd events from ns kuttl-test-helped-vulture: logger.go:42: 07:17:47 | examples-openshift-with-htpasswd | 2024-01-08 07:17:33 +0000 UTC Normal Pod with-htpasswd-79fd8fd57c-bqrrt Binding Scheduled Successfully assigned kuttl-test-helped-vulture/with-htpasswd-79fd8fd57c-bqrrt to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:17:47 | examples-openshift-with-htpasswd | 2024-01-08 07:17:33 +0000 UTC Normal ReplicaSet.apps with-htpasswd-79fd8fd57c SuccessfulCreate Created pod: with-htpasswd-79fd8fd57c-bqrrt replicaset-controller logger.go:42: 07:17:47 | examples-openshift-with-htpasswd | 2024-01-08 07:17:33 +0000 UTC Normal Deployment.apps with-htpasswd ScalingReplicaSet Scaled up replica set with-htpasswd-79fd8fd57c to 1 deployment-controller logger.go:42: 07:17:47 | examples-openshift-with-htpasswd | 2024-01-08 07:17:34 +0000 UTC Normal Pod with-htpasswd-79fd8fd57c-bqrrt AddedInterface Add eth0 [10.131.0.51/23] from ovn-kubernetes logger.go:42: 07:17:47 | examples-openshift-with-htpasswd | 2024-01-08 07:17:34 +0000 UTC Normal Pod with-htpasswd-79fd8fd57c-bqrrt.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:17:47 | examples-openshift-with-htpasswd | 2024-01-08 07:17:34 +0000 UTC Normal Pod with-htpasswd-79fd8fd57c-bqrrt.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:17:47 | examples-openshift-with-htpasswd | 2024-01-08 07:17:34 +0000 UTC Normal Pod with-htpasswd-79fd8fd57c-bqrrt.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:17:47 | examples-openshift-with-htpasswd | 2024-01-08 07:17:34 +0000 UTC Normal Pod with-htpasswd-79fd8fd57c-bqrrt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:17:47 | examples-openshift-with-htpasswd | 2024-01-08 07:17:34 +0000 UTC Normal Pod with-htpasswd-79fd8fd57c-bqrrt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:17:47 | examples-openshift-with-htpasswd | 2024-01-08 07:17:34 +0000 UTC Normal Pod with-htpasswd-79fd8fd57c-bqrrt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:17:48 | examples-openshift-with-htpasswd | Deleting namespace: kuttl-test-helped-vulture === CONT kuttl/harness/examples-service-types logger.go:42: 07:17:54 | examples-service-types | Creating namespace: kuttl-test-present-mammal logger.go:42: 07:17:54 | examples-service-types/0-install | starting test step 0-install logger.go:42: 07:17:54 | examples-service-types/0-install | Jaeger:kuttl-test-present-mammal/service-types created logger.go:42: 07:17:59 | examples-service-types/0-install | test step completed 0-install logger.go:42: 07:17:59 | examples-service-types/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:17:59 | examples-service-types/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE service-types /dev/null] logger.go:42: 07:18:01 | examples-service-types/1-smoke-test | Warning: resource jaegers/service-types is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:18:07 | examples-service-types/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://service-types-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:18:08 | examples-service-types/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:18:08 | examples-service-types/1-smoke-test | job.batch/report-span created logger.go:42: 07:18:08 | examples-service-types/1-smoke-test | job.batch/check-span created logger.go:42: 07:18:19 | examples-service-types/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:18:19 | examples-service-types/2- | starting test step 2- logger.go:42: 07:18:19 | examples-service-types/2- | test step completed 2- logger.go:42: 07:18:20 | examples-service-types | examples-service-types events from ns kuttl-test-present-mammal: logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:17:57 +0000 UTC Normal Service service-types-query EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:17:58 +0000 UTC Normal Pod service-types-78cf64f9cf-6ttnl Binding Scheduled Successfully assigned kuttl-test-present-mammal/service-types-78cf64f9cf-6ttnl to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:17:58 +0000 UTC Normal Pod service-types-78cf64f9cf-6ttnl AddedInterface Add eth0 [10.128.2.67/23] from ovn-kubernetes logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:17:58 +0000 UTC Normal Pod service-types-78cf64f9cf-6ttnl.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:17:58 +0000 UTC Normal Pod service-types-78cf64f9cf-6ttnl.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:17:58 +0000 UTC Normal Pod service-types-78cf64f9cf-6ttnl.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:17:58 +0000 UTC Normal Pod service-types-78cf64f9cf-6ttnl.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:17:58 +0000 UTC Normal Pod service-types-78cf64f9cf-6ttnl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:17:58 +0000 UTC Normal Pod service-types-78cf64f9cf-6ttnl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:17:58 +0000 UTC Normal ReplicaSet.apps service-types-78cf64f9cf SuccessfulCreate Created pod: service-types-78cf64f9cf-6ttnl replicaset-controller logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:17:58 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-78cf64f9cf to 1 deployment-controller logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:01 +0000 UTC Normal Service service-types-collector EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:01 +0000 UTC Normal Service service-types-query EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:04 +0000 UTC Normal Service service-types-collector EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:05 +0000 UTC Normal Pod service-types-78cf64f9cf-6ttnl.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:05 +0000 UTC Normal Pod service-types-78cf64f9cf-6ttnl.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:05 +0000 UTC Normal ReplicaSet.apps service-types-78cf64f9cf SuccessfulDelete Deleted pod: service-types-78cf64f9cf-6ttnl replicaset-controller logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:05 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled down replica set service-types-78cf64f9cf to 0 from 1 deployment-controller logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:06 +0000 UTC Normal Pod service-types-7d8dcb6d77-bbp9g Binding Scheduled Successfully assigned kuttl-test-present-mammal/service-types-7d8dcb6d77-bbp9g to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:06 +0000 UTC Normal Pod service-types-7d8dcb6d77-bbp9g AddedInterface Add eth0 [10.128.2.68/23] from ovn-kubernetes logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:06 +0000 UTC Normal Pod service-types-7d8dcb6d77-bbp9g.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:06 +0000 UTC Normal Pod service-types-7d8dcb6d77-bbp9g.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:06 +0000 UTC Normal Pod service-types-7d8dcb6d77-bbp9g.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:06 +0000 UTC Normal Pod service-types-7d8dcb6d77-bbp9g.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:06 +0000 UTC Normal ReplicaSet.apps service-types-7d8dcb6d77 SuccessfulCreate Created pod: service-types-7d8dcb6d77-bbp9g replicaset-controller logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:06 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-7d8dcb6d77 to 1 deployment-controller logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:07 +0000 UTC Normal Pod service-types-7d8dcb6d77-bbp9g.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:07 +0000 UTC Normal Pod service-types-7d8dcb6d77-bbp9g.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:08 +0000 UTC Normal Pod check-span-6smc2 Binding Scheduled Successfully assigned kuttl-test-present-mammal/check-span-6smc2 to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:08 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-6smc2 job-controller logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:08 +0000 UTC Normal Pod report-span-zhthm Binding Scheduled Successfully assigned kuttl-test-present-mammal/report-span-zhthm to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:08 +0000 UTC Normal Pod report-span-zhthm AddedInterface Add eth0 [10.129.2.37/23] from ovn-kubernetes logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:08 +0000 UTC Normal Pod report-span-zhthm.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:08 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-zhthm job-controller logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:09 +0000 UTC Normal Pod check-span-6smc2 AddedInterface Add eth0 [10.131.0.52/23] from ovn-kubernetes logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:09 +0000 UTC Normal Pod check-span-6smc2.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:09 +0000 UTC Normal Pod check-span-6smc2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:09 +0000 UTC Normal Pod check-span-6smc2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:09 +0000 UTC Normal Pod report-span-zhthm.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:09 +0000 UTC Normal Pod report-span-zhthm.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:18:20 | examples-service-types | 2024-01-08 07:18:19 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:18:20 | examples-service-types | Deleting namespace: kuttl-test-present-mammal === CONT kuttl/harness/examples-collector-with-priority-class logger.go:42: 07:18:48 | examples-collector-with-priority-class | Creating namespace: kuttl-test-splendid-dolphin logger.go:42: 07:18:48 | examples-collector-with-priority-class/0-install | starting test step 0-install logger.go:42: 07:18:48 | examples-collector-with-priority-class/0-install | PriorityClass:/collector-high-priority created logger.go:42: 07:18:48 | examples-collector-with-priority-class/0-install | Jaeger:kuttl-test-splendid-dolphin/collector-with-high-priority created logger.go:42: 07:18:54 | examples-collector-with-priority-class/0-install | test step completed 0-install logger.go:42: 07:18:54 | examples-collector-with-priority-class/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:18:54 | examples-collector-with-priority-class/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE collector-with-high-priority /dev/null] logger.go:42: 07:18:56 | examples-collector-with-priority-class/1-smoke-test | Warning: resource jaegers/collector-with-high-priority is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:19:03 | examples-collector-with-priority-class/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:19:03 | examples-collector-with-priority-class/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:19:04 | examples-collector-with-priority-class/1-smoke-test | job.batch/report-span created logger.go:42: 07:19:04 | examples-collector-with-priority-class/1-smoke-test | job.batch/check-span created logger.go:42: 07:29:04 | examples-collector-with-priority-class/1-smoke-test | test step failed 1-smoke-test case.go:364: failed in step 1-smoke-test case.go:366: --- Job:kuttl-test-splendid-dolphin/check-span +++ Job:kuttl-test-splendid-dolphin/check-span @@ -1,8 +1,141 @@ apiVersion: batch/v1 kind: Job metadata: + annotations: + kubectl.kubernetes.io/last-applied-configuration: | + {"apiVersion":"batch/v1","kind":"Job","metadata":{"annotations":{},"name":"check-span","namespace":"kuttl-test-splendid-dolphin"},"spec":{"backoffLimit":15,"template":{"spec":{"containers":[{"command":["./query"],"env":[{"name":"SERVICE_NAME","value":"smoke-test-service"},{"name":"QUERY_HOST","value":"https://collector-with-high-priority-query:443"},{"name":"SECRET_PATH","value":"/var/run/secrets/api-token/token"}],"image":"registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344","name":"asserts-container","volumeMounts":[{"mountPath":"/var/run/secrets/api-token","name":"token-api-volume"}]}],"restartPolicy":"OnFailure","volumes":[{"name":"token-api-volume","secret":{"secretName":"e2e-test"}}]}}}} + labels: + batch.kubernetes.io/controller-uid: 6922f232-2910-4e2e-b3ae-b0148aa88785 + batch.kubernetes.io/job-name: check-span + controller-uid: 6922f232-2910-4e2e-b3ae-b0148aa88785 + job-name: check-span + managedFields: + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:kubectl.kubernetes.io/last-applied-configuration: {} + f:spec: + f:backoffLimit: {} + f:completionMode: {} + f:completions: {} + f:parallelism: {} + f:suspend: {} + f:template: + f:spec: + f:containers: + k:{"name":"asserts-container"}: + .: {} + f:command: {} + f:env: + .: {} + k:{"name":"QUERY_HOST"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SECRET_PATH"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SERVICE_NAME"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:name: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/var/run/secrets/api-token"}: + .: {} + f:mountPath: {} + f:name: {} + f:dnsPolicy: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"token-api-volume"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + manager: kubectl-client-side-apply + operation: Update + time: "2024-01-08T07:19:04Z" + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:status: + f:active: {} + f:ready: {} + f:startTime: {} + f:uncountedTerminatedPods: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-01-08T07:27:50Z" name: check-span namespace: kuttl-test-splendid-dolphin +spec: + backoffLimit: 15 + completionMode: NonIndexed + completions: 1 + parallelism: 1 + selector: + matchLabels: + batch.kubernetes.io/controller-uid: 6922f232-2910-4e2e-b3ae-b0148aa88785 + suspend: false + template: + metadata: + creationTimestamp: null + labels: + batch.kubernetes.io/controller-uid: 6922f232-2910-4e2e-b3ae-b0148aa88785 + batch.kubernetes.io/job-name: check-span + controller-uid: 6922f232-2910-4e2e-b3ae-b0148aa88785 + job-name: check-span + spec: + containers: + - command: + - ./query + env: + - name: SERVICE_NAME + value: smoke-test-service + - name: QUERY_HOST + value: https://collector-with-high-priority-query:443 + - name: SECRET_PATH + value: /var/run/secrets/api-token/token + image: registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 + imagePullPolicy: IfNotPresent + name: asserts-container + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /var/run/secrets/api-token + name: token-api-volume + dnsPolicy: ClusterFirst + restartPolicy: OnFailure + schedulerName: default-scheduler + securityContext: {} + terminationGracePeriodSeconds: 30 + volumes: + - name: token-api-volume + secret: + defaultMode: 420 + secretName: e2e-test status: - succeeded: 1 + active: 1 + ready: 1 + startTime: "2024-01-08T07:19:04Z" + uncountedTerminatedPods: {} case.go:366: resource Job:kuttl-test-splendid-dolphin/check-span: .status.succeeded: key is missing from map logger.go:42: 07:29:04 | examples-collector-with-priority-class | examples-collector-with-priority-class events from ns kuttl-test-splendid-dolphin: logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:18:51 +0000 UTC Normal Pod collector-with-high-priority-6697ddc76b-s2bg4 Binding Scheduled Successfully assigned kuttl-test-splendid-dolphin/collector-with-high-priority-6697ddc76b-s2bg4 to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:18:51 +0000 UTC Normal ReplicaSet.apps collector-with-high-priority-6697ddc76b SuccessfulCreate Created pod: collector-with-high-priority-6697ddc76b-s2bg4 replicaset-controller logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:18:51 +0000 UTC Normal Deployment.apps collector-with-high-priority ScalingReplicaSet Scaled up replica set collector-with-high-priority-6697ddc76b to 1 deployment-controller logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:18:52 +0000 UTC Warning Pod collector-with-high-priority-6697ddc76b-s2bg4 FailedMount MountVolume.SetUp failed for volume "collector-with-high-priority-ui-oauth-proxy-tls" : secret "collector-with-high-priority-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:18:52 +0000 UTC Warning Pod collector-with-high-priority-6697ddc76b-s2bg4 FailedMount MountVolume.SetUp failed for volume "collector-with-high-priority-collector-tls-config-volume" : secret "collector-with-high-priority-collector-headless-tls" not found kubelet logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:18:52 +0000 UTC Normal Pod collector-with-high-priority-6697ddc76b-s2bg4 AddedInterface Add eth0 [10.131.0.53/23] from ovn-kubernetes logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:18:52 +0000 UTC Normal Pod collector-with-high-priority-6697ddc76b-s2bg4.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:18:53 +0000 UTC Normal Pod collector-with-high-priority-6697ddc76b-s2bg4.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:18:53 +0000 UTC Normal Pod collector-with-high-priority-6697ddc76b-s2bg4.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:18:53 +0000 UTC Normal Pod collector-with-high-priority-6697ddc76b-s2bg4.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:18:53 +0000 UTC Normal Pod collector-with-high-priority-6697ddc76b-s2bg4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:18:53 +0000 UTC Normal Pod collector-with-high-priority-6697ddc76b-s2bg4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:19:04 +0000 UTC Normal Pod check-span-fmbgw Binding Scheduled Successfully assigned kuttl-test-splendid-dolphin/check-span-fmbgw to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:19:04 +0000 UTC Normal Pod check-span-fmbgw AddedInterface Add eth0 [10.128.2.69/23] from ovn-kubernetes logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:19:04 +0000 UTC Normal Pod check-span-fmbgw.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:19:04 +0000 UTC Normal Pod check-span-fmbgw.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:19:04 +0000 UTC Normal Pod check-span-fmbgw.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:19:04 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-fmbgw job-controller logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:19:04 +0000 UTC Normal Pod report-span-qsz7q Binding Scheduled Successfully assigned kuttl-test-splendid-dolphin/report-span-qsz7q to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:19:04 +0000 UTC Normal Pod report-span-qsz7q AddedInterface Add eth0 [10.129.2.38/23] from ovn-kubernetes logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:19:04 +0000 UTC Normal Pod report-span-qsz7q.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:19:04 +0000 UTC Normal Pod report-span-qsz7q.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:19:04 +0000 UTC Normal Pod report-span-qsz7q.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:19:04 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-qsz7q job-controller logger.go:42: 07:29:04 | examples-collector-with-priority-class | 2024-01-08 07:24:26 +0000 UTC Warning Pod check-span-fmbgw.spec.containers{asserts-container} BackOff Back-off restarting failed container asserts-container in pod check-span-fmbgw_kuttl-test-splendid-dolphin(6f67ebfe-c54d-45c5-b375-92cf8fcd1bb3) kubelet logger.go:42: 07:29:04 | examples-collector-with-priority-class | Deleting namespace: kuttl-test-splendid-dolphin === CONT kuttl/harness/examples-agent-with-priority-class logger.go:42: 07:29:17 | examples-agent-with-priority-class | Creating namespace: kuttl-test-pleasing-gull logger.go:42: 07:29:17 | examples-agent-with-priority-class/0-install | starting test step 0-install logger.go:42: 07:29:17 | examples-agent-with-priority-class/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 07:29:17 | examples-agent-with-priority-class/0-install | ServiceAccount:kuttl-test-pleasing-gull/jaeger-agent-daemonset created logger.go:42: 07:29:17 | examples-agent-with-priority-class/0-install | test step completed 0-install logger.go:42: 07:29:17 | examples-agent-with-priority-class/1-add-policy | starting test step 1-add-policy logger.go:42: 07:29:17 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 07:29:17 | examples-agent-with-priority-class/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 07:29:17 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 07:29:22 | examples-agent-with-priority-class/1-add-policy | test step completed 1-add-policy logger.go:42: 07:29:22 | examples-agent-with-priority-class/2-install | starting test step 2-install logger.go:42: 07:29:22 | examples-agent-with-priority-class/2-install | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 07:29:31 | examples-agent-with-priority-class/2-install | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:29:31 | examples-agent-with-priority-class/2-install | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:29:32 | examples-agent-with-priority-class/2-install | job.batch/report-span created logger.go:42: 07:29:32 | examples-agent-with-priority-class/2-install | job.batch/check-span created logger.go:42: 07:29:32 | examples-agent-with-priority-class/2-install | PriorityClass:/high-priority created logger.go:42: 07:29:32 | examples-agent-with-priority-class/2-install | Jaeger:kuttl-test-pleasing-gull/agent-as-daemonset updated logger.go:42: 07:29:44 | examples-agent-with-priority-class/2-install | test step completed 2-install logger.go:42: 07:29:44 | examples-agent-with-priority-class | examples-agent-with-priority-class events from ns kuttl-test-pleasing-gull: logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:28 +0000 UTC Normal Pod agent-as-daemonset-6967cf6b58-bznhj Binding Scheduled Successfully assigned kuttl-test-pleasing-gull/agent-as-daemonset-6967cf6b58-bznhj to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:28 +0000 UTC Normal Pod agent-as-daemonset-6967cf6b58-bznhj AddedInterface Add eth0 [10.128.2.70/23] from ovn-kubernetes logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:28 +0000 UTC Normal Pod agent-as-daemonset-6967cf6b58-bznhj.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:28 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-6967cf6b58 SuccessfulCreate Created pod: agent-as-daemonset-6967cf6b58-bznhj replicaset-controller logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:28 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-6967cf6b58 to 1 deployment-controller logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:29 +0000 UTC Normal Pod agent-as-daemonset-6967cf6b58-bznhj.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:29 +0000 UTC Normal Pod agent-as-daemonset-6967cf6b58-bznhj.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:29 +0000 UTC Normal Pod agent-as-daemonset-6967cf6b58-bznhj.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:29 +0000 UTC Normal Pod agent-as-daemonset-6967cf6b58-bznhj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:29 +0000 UTC Normal Pod agent-as-daemonset-6967cf6b58-bznhj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:32 +0000 UTC Normal Pod check-span-tdbkz Binding Scheduled Successfully assigned kuttl-test-pleasing-gull/check-span-tdbkz to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:32 +0000 UTC Normal Pod check-span-tdbkz AddedInterface Add eth0 [10.131.0.54/23] from ovn-kubernetes logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:32 +0000 UTC Normal Pod check-span-tdbkz.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:32 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-tdbkz job-controller logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:32 +0000 UTC Normal Pod report-span-ql5f2 Binding Scheduled Successfully assigned kuttl-test-pleasing-gull/report-span-ql5f2 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:32 +0000 UTC Normal Pod report-span-ql5f2 AddedInterface Add eth0 [10.129.2.39/23] from ovn-kubernetes logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:32 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-ql5f2 job-controller logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:33 +0000 UTC Normal Pod check-span-tdbkz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:33 +0000 UTC Normal Pod check-span-tdbkz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:33 +0000 UTC Normal Pod report-span-ql5f2.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:33 +0000 UTC Normal Pod report-span-ql5f2.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:33 +0000 UTC Normal Pod report-span-ql5f2.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:35 +0000 UTC Warning DaemonSet.apps agent-as-daemonset-agent-daemonset FailedCreate Error creating: pods "agent-as-daemonset-agent-daemonset-" is forbidden: unable to validate against any security context constraint: [provider "anyuid": Forbidden: not usable by user or serviceaccount, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5775: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5778: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6831: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6832: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 14271: Host ports are not allowed to be used, provider "restricted": Forbidden: not usable by user or serviceaccount, provider "nonroot-v2": Forbidden: not usable by user or serviceaccount, provider "nonroot": Forbidden: not usable by user or serviceaccount, provider "hostmount-anyuid": Forbidden: not usable by user or serviceaccount, provider "elasticsearch-scc": Forbidden: not usable by user or serviceaccount, provider "machine-api-termination-handler": Forbidden: not usable by user or serviceaccount, provider "daemonset-with-hostport": Forbidden: not usable by user or serviceaccount, provider "hostnetwork-v2": Forbidden: not usable by user or serviceaccount, provider "hostnetwork": Forbidden: not usable by user or serviceaccount, provider "hostaccess": Forbidden: not usable by user or serviceaccount, provider "node-exporter": Forbidden: not usable by user or serviceaccount, provider "privileged": Forbidden: not usable by user or serviceaccount] daemonset-controller logger.go:42: 07:29:44 | examples-agent-with-priority-class | 2024-01-08 07:29:43 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:29:44 | examples-agent-with-priority-class | Deleting namespace: kuttl-test-pleasing-gull === CONT kuttl/harness/examples-all-in-one-with-options logger.go:42: 07:29:57 | examples-all-in-one-with-options | Creating namespace: kuttl-test-prompt-bat logger.go:42: 07:29:57 | examples-all-in-one-with-options/0-install | starting test step 0-install logger.go:42: 07:29:57 | examples-all-in-one-with-options/0-install | Jaeger:kuttl-test-prompt-bat/my-jaeger created logger.go:42: 07:30:04 | examples-all-in-one-with-options/0-install | test step completed 0-install logger.go:42: 07:30:04 | examples-all-in-one-with-options/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:30:04 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:30:06 | examples-all-in-one-with-options/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:30:13 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443/jaeger MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:30:13 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:30:14 | examples-all-in-one-with-options/1-smoke-test | job.batch/report-span created logger.go:42: 07:30:14 | examples-all-in-one-with-options/1-smoke-test | job.batch/check-span created logger.go:42: 07:30:25 | examples-all-in-one-with-options/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:30:25 | examples-all-in-one-with-options | examples-all-in-one-with-options events from ns kuttl-test-prompt-bat: logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:01 +0000 UTC Normal Pod my-jaeger-7f9bbdc8-skb26 Binding Scheduled Successfully assigned kuttl-test-prompt-bat/my-jaeger-7f9bbdc8-skb26 to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:01 +0000 UTC Normal Pod my-jaeger-7f9bbdc8-skb26 AddedInterface Add eth0 [10.131.0.55/23] from ovn-kubernetes logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:01 +0000 UTC Normal ReplicaSet.apps my-jaeger-7f9bbdc8 SuccessfulCreate Created pod: my-jaeger-7f9bbdc8-skb26 replicaset-controller logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:01 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-7f9bbdc8 to 1 deployment-controller logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:02 +0000 UTC Normal Pod my-jaeger-7f9bbdc8-skb26.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:02 +0000 UTC Normal Pod my-jaeger-7f9bbdc8-skb26.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:02 +0000 UTC Normal Pod my-jaeger-7f9bbdc8-skb26.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:02 +0000 UTC Normal Pod my-jaeger-7f9bbdc8-skb26.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:02 +0000 UTC Normal Pod my-jaeger-7f9bbdc8-skb26.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:02 +0000 UTC Normal Pod my-jaeger-7f9bbdc8-skb26.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:07 +0000 UTC Normal Pod my-jaeger-7f9bbdc8-skb26.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:07 +0000 UTC Normal Pod my-jaeger-7f9bbdc8-skb26.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:07 +0000 UTC Normal ReplicaSet.apps my-jaeger-7f9bbdc8 SuccessfulDelete Deleted pod: my-jaeger-7f9bbdc8-skb26 replicaset-controller logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:07 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-7f9bbdc8 to 0 from 1 deployment-controller logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:09 +0000 UTC Normal Pod my-jaeger-6649d94f5f-6k58h Binding Scheduled Successfully assigned kuttl-test-prompt-bat/my-jaeger-6649d94f5f-6k58h to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:09 +0000 UTC Normal Pod my-jaeger-6649d94f5f-6k58h AddedInterface Add eth0 [10.131.0.56/23] from ovn-kubernetes logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:09 +0000 UTC Normal Pod my-jaeger-6649d94f5f-6k58h.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:09 +0000 UTC Normal Pod my-jaeger-6649d94f5f-6k58h.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:09 +0000 UTC Normal Pod my-jaeger-6649d94f5f-6k58h.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:09 +0000 UTC Normal Pod my-jaeger-6649d94f5f-6k58h.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:09 +0000 UTC Normal Pod my-jaeger-6649d94f5f-6k58h.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:09 +0000 UTC Normal Pod my-jaeger-6649d94f5f-6k58h.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:09 +0000 UTC Normal ReplicaSet.apps my-jaeger-6649d94f5f SuccessfulCreate Created pod: my-jaeger-6649d94f5f-6k58h replicaset-controller logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:09 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6649d94f5f to 1 deployment-controller logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:14 +0000 UTC Normal Pod check-span-fcd87 Binding Scheduled Successfully assigned kuttl-test-prompt-bat/check-span-fcd87 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:14 +0000 UTC Normal Pod check-span-fcd87 AddedInterface Add eth0 [10.128.2.71/23] from ovn-kubernetes logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:14 +0000 UTC Normal Pod check-span-fcd87.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:14 +0000 UTC Normal Pod check-span-fcd87.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:14 +0000 UTC Normal Pod check-span-fcd87.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:14 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-fcd87 job-controller logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:14 +0000 UTC Normal Pod report-span-68wgj Binding Scheduled Successfully assigned kuttl-test-prompt-bat/report-span-68wgj to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:14 +0000 UTC Normal Pod report-span-68wgj AddedInterface Add eth0 [10.129.2.41/23] from ovn-kubernetes logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:14 +0000 UTC Normal Pod report-span-68wgj.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:14 +0000 UTC Normal Pod report-span-68wgj.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:14 +0000 UTC Normal Pod report-span-68wgj.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:14 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-68wgj job-controller logger.go:42: 07:30:25 | examples-all-in-one-with-options | 2024-01-08 07:30:25 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:30:25 | examples-all-in-one-with-options | Deleting namespace: kuttl-test-prompt-bat === CONT kuttl/harness/examples-agent-as-daemonset logger.go:42: 07:30:38 | examples-agent-as-daemonset | Creating namespace: kuttl-test-generous-martin logger.go:42: 07:30:38 | examples-agent-as-daemonset/0-install | starting test step 0-install logger.go:42: 07:30:38 | examples-agent-as-daemonset/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 07:30:38 | examples-agent-as-daemonset/0-install | ServiceAccount:kuttl-test-generous-martin/jaeger-agent-daemonset created logger.go:42: 07:30:38 | examples-agent-as-daemonset/0-install | test step completed 0-install logger.go:42: 07:30:38 | examples-agent-as-daemonset/1-add-policy | starting test step 1-add-policy logger.go:42: 07:30:38 | examples-agent-as-daemonset/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 07:30:38 | examples-agent-as-daemonset/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 07:30:38 | examples-agent-as-daemonset/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 07:30:43 | examples-agent-as-daemonset/1-add-policy | test step completed 1-add-policy logger.go:42: 07:30:43 | examples-agent-as-daemonset/2-install | starting test step 2-install logger.go:42: 07:30:44 | examples-agent-as-daemonset/2-install | Jaeger:kuttl-test-generous-martin/agent-as-daemonset created logger.go:42: 07:30:44 | examples-agent-as-daemonset/2-install | test step completed 2-install logger.go:42: 07:30:44 | examples-agent-as-daemonset/3- | starting test step 3- logger.go:42: 07:30:49 | examples-agent-as-daemonset/3- | test step completed 3- logger.go:42: 07:30:49 | examples-agent-as-daemonset | examples-agent-as-daemonset events from ns kuttl-test-generous-martin: logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:47 +0000 UTC Normal Pod agent-as-daemonset-75b6f6db7b-ppl9r Binding Scheduled Successfully assigned kuttl-test-generous-martin/agent-as-daemonset-75b6f6db7b-ppl9r to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:47 +0000 UTC Normal Pod agent-as-daemonset-75b6f6db7b-ppl9r AddedInterface Add eth0 [10.131.0.57/23] from ovn-kubernetes logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:47 +0000 UTC Normal Pod agent-as-daemonset-75b6f6db7b-ppl9r.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:47 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-75b6f6db7b SuccessfulCreate Created pod: agent-as-daemonset-75b6f6db7b-ppl9r replicaset-controller logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:47 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-dchzd Binding Scheduled Successfully assigned kuttl-test-generous-martin/agent-as-daemonset-agent-daemonset-dchzd to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:47 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-qm9t2 Binding Scheduled Successfully assigned kuttl-test-generous-martin/agent-as-daemonset-agent-daemonset-qm9t2 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:47 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-wr9qw Binding Scheduled Successfully assigned kuttl-test-generous-martin/agent-as-daemonset-agent-daemonset-wr9qw to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:47 +0000 UTC Normal DaemonSet.apps agent-as-daemonset-agent-daemonset SuccessfulCreate Created pod: agent-as-daemonset-agent-daemonset-wr9qw daemonset-controller logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:47 +0000 UTC Normal DaemonSet.apps agent-as-daemonset-agent-daemonset SuccessfulCreate Created pod: agent-as-daemonset-agent-daemonset-dchzd daemonset-controller logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:47 +0000 UTC Normal DaemonSet.apps agent-as-daemonset-agent-daemonset SuccessfulCreate Created pod: agent-as-daemonset-agent-daemonset-qm9t2 daemonset-controller logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:47 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-75b6f6db7b to 1 deployment-controller logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:48 +0000 UTC Normal Pod agent-as-daemonset-75b6f6db7b-ppl9r.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:48 +0000 UTC Normal Pod agent-as-daemonset-75b6f6db7b-ppl9r.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:48 +0000 UTC Normal Pod agent-as-daemonset-75b6f6db7b-ppl9r.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:48 +0000 UTC Normal Pod agent-as-daemonset-75b6f6db7b-ppl9r.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:48 +0000 UTC Normal Pod agent-as-daemonset-75b6f6db7b-ppl9r.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:48 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-dchzd AddedInterface Add eth0 [10.131.0.58/23] from ovn-kubernetes logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:48 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-dchzd.spec.containers{jaeger-agent-daemonset} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:48 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-dchzd.spec.containers{jaeger-agent-daemonset} Created Created container jaeger-agent-daemonset kubelet logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:48 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-dchzd.spec.containers{jaeger-agent-daemonset} Started Started container jaeger-agent-daemonset kubelet logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:48 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-qm9t2 AddedInterface Add eth0 [10.129.2.42/23] from ovn-kubernetes logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:48 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-qm9t2.spec.containers{jaeger-agent-daemonset} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" kubelet logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:48 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-wr9qw AddedInterface Add eth0 [10.128.2.72/23] from ovn-kubernetes logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:48 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-wr9qw.spec.containers{jaeger-agent-daemonset} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:48 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-wr9qw.spec.containers{jaeger-agent-daemonset} Created Created container jaeger-agent-daemonset kubelet logger.go:42: 07:30:49 | examples-agent-as-daemonset | 2024-01-08 07:30:48 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-wr9qw.spec.containers{jaeger-agent-daemonset} Started Started container jaeger-agent-daemonset kubelet logger.go:42: 07:30:49 | examples-agent-as-daemonset | Deleting namespace: kuttl-test-generous-martin === CONT kuttl/harness/examples-business-application-injected-sidecar logger.go:42: 07:30:56 | examples-business-application-injected-sidecar | Creating namespace: kuttl-test-inviting-zebra logger.go:42: 07:30:56 | examples-business-application-injected-sidecar/0-install | starting test step 0-install logger.go:42: 07:30:56 | examples-business-application-injected-sidecar/0-install | Deployment:kuttl-test-inviting-zebra/myapp created logger.go:42: 07:30:56 | examples-business-application-injected-sidecar/0-install | test step completed 0-install logger.go:42: 07:30:56 | examples-business-application-injected-sidecar/1-install | starting test step 1-install logger.go:42: 07:30:56 | examples-business-application-injected-sidecar/1-install | Jaeger:kuttl-test-inviting-zebra/simplest created logger.go:42: 07:31:06 | examples-business-application-injected-sidecar/1-install | test step completed 1-install logger.go:42: 07:31:06 | examples-business-application-injected-sidecar/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:31:06 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:31:08 | examples-business-application-injected-sidecar/2-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:31:14 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:31:15 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:31:15 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/report-span created logger.go:42: 07:31:15 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/check-span created logger.go:42: 07:31:26 | examples-business-application-injected-sidecar/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | examples-business-application-injected-sidecar events from ns kuttl-test-inviting-zebra: logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:30:56 +0000 UTC Normal Pod myapp-679f79d5f8-tgzjs Binding Scheduled Successfully assigned kuttl-test-inviting-zebra/myapp-679f79d5f8-tgzjs to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:30:56 +0000 UTC Normal Pod myapp-679f79d5f8-tgzjs AddedInterface Add eth0 [10.129.2.43/23] from ovn-kubernetes logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:30:56 +0000 UTC Normal Pod myapp-679f79d5f8-tgzjs.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:30:56 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulCreate Created pod: myapp-679f79d5f8-tgzjs replicaset-controller logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:30:56 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-679f79d5f8 to 1 deployment-controller logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:30:57 +0000 UTC Normal Pod myapp-78ddfc878d-pmnwq Binding Scheduled Successfully assigned kuttl-test-inviting-zebra/myapp-78ddfc878d-pmnwq to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:30:57 +0000 UTC Warning Pod myapp-78ddfc878d-pmnwq FailedMount MountVolume.SetUp failed for volume "simplest-service-ca" : configmap "simplest-service-ca" not found kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:30:57 +0000 UTC Warning Pod myapp-78ddfc878d-pmnwq FailedMount MountVolume.SetUp failed for volume "simplest-trusted-ca" : configmap "simplest-trusted-ca" not found kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:30:57 +0000 UTC Normal ReplicaSet.apps myapp-78ddfc878d SuccessfulCreate Created pod: myapp-78ddfc878d-pmnwq replicaset-controller logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:30:57 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-78ddfc878d to 1 deployment-controller logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:00 +0000 UTC Normal Pod myapp-679f79d5f8-tgzjs.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.769s (3.769s including waiting) kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:00 +0000 UTC Normal Pod myapp-679f79d5f8-tgzjs.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:00 +0000 UTC Normal Pod myapp-679f79d5f8-tgzjs.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:04 +0000 UTC Normal Pod simplest-894c7b769-c99gc Binding Scheduled Successfully assigned kuttl-test-inviting-zebra/simplest-894c7b769-c99gc to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:04 +0000 UTC Warning Pod simplest-894c7b769-c99gc FailedMount MountVolume.SetUp failed for volume "simplest-collector-tls-config-volume" : secret "simplest-collector-headless-tls" not found kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:04 +0000 UTC Normal ReplicaSet.apps simplest-894c7b769 SuccessfulCreate Created pod: simplest-894c7b769-c99gc replicaset-controller logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:04 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-894c7b769 to 1 deployment-controller logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:05 +0000 UTC Normal Pod myapp-78ddfc878d-pmnwq AddedInterface Add eth0 [10.128.2.73/23] from ovn-kubernetes logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:05 +0000 UTC Normal Pod myapp-78ddfc878d-pmnwq.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:05 +0000 UTC Normal Pod simplest-894c7b769-c99gc AddedInterface Add eth0 [10.131.0.59/23] from ovn-kubernetes logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:05 +0000 UTC Normal Pod simplest-894c7b769-c99gc.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:05 +0000 UTC Normal Pod simplest-894c7b769-c99gc.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:05 +0000 UTC Normal Pod simplest-894c7b769-c99gc.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:05 +0000 UTC Normal Pod simplest-894c7b769-c99gc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:05 +0000 UTC Normal Pod simplest-894c7b769-c99gc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:05 +0000 UTC Normal Pod simplest-894c7b769-c99gc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:07 +0000 UTC Warning Pod myapp-679f79d5f8-tgzjs.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.129.2.43:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:10 +0000 UTC Normal Pod myapp-78ddfc878d-pmnwq.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 5.107s (5.107s including waiting) kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:10 +0000 UTC Normal Pod myapp-78ddfc878d-pmnwq.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:10 +0000 UTC Normal Pod myapp-78ddfc878d-pmnwq.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:10 +0000 UTC Normal Pod myapp-78ddfc878d-pmnwq.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:10 +0000 UTC Normal Pod myapp-78ddfc878d-pmnwq.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:10 +0000 UTC Normal Pod myapp-78ddfc878d-pmnwq.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:10 +0000 UTC Normal Pod simplest-894c7b769-c99gc.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:10 +0000 UTC Normal Pod simplest-894c7b769-c99gc.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:10 +0000 UTC Normal ReplicaSet.apps simplest-894c7b769 SuccessfulDelete Deleted pod: simplest-894c7b769-c99gc replicaset-controller logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:10 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-894c7b769 to 0 from 1 deployment-controller logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:11 +0000 UTC Normal Pod simplest-9c6c9cc75-shbrs Binding Scheduled Successfully assigned kuttl-test-inviting-zebra/simplest-9c6c9cc75-shbrs to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:11 +0000 UTC Normal Pod simplest-9c6c9cc75-shbrs AddedInterface Add eth0 [10.131.0.60/23] from ovn-kubernetes logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:11 +0000 UTC Normal Pod simplest-9c6c9cc75-shbrs.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:11 +0000 UTC Normal Pod simplest-9c6c9cc75-shbrs.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:11 +0000 UTC Normal Pod simplest-9c6c9cc75-shbrs.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:11 +0000 UTC Normal Pod simplest-9c6c9cc75-shbrs.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:11 +0000 UTC Normal Pod simplest-9c6c9cc75-shbrs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:11 +0000 UTC Normal Pod simplest-9c6c9cc75-shbrs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:11 +0000 UTC Normal ReplicaSet.apps simplest-9c6c9cc75 SuccessfulCreate Created pod: simplest-9c6c9cc75-shbrs replicaset-controller logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:11 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-9c6c9cc75 to 1 deployment-controller logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:12 +0000 UTC Normal Pod myapp-679f79d5f8-tgzjs.spec.containers{myapp} Killing Stopping container myapp kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:12 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulDelete Deleted pod: myapp-679f79d5f8-tgzjs replicaset-controller logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:12 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled down replica set myapp-679f79d5f8 to 0 from 1 deployment-controller logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:15 +0000 UTC Normal Pod check-span-55pn5 Binding Scheduled Successfully assigned kuttl-test-inviting-zebra/check-span-55pn5 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:15 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-55pn5 job-controller logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:15 +0000 UTC Normal Pod report-span-v7sgp Binding Scheduled Successfully assigned kuttl-test-inviting-zebra/report-span-v7sgp to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:15 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-v7sgp job-controller logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:16 +0000 UTC Normal Pod check-span-55pn5 AddedInterface Add eth0 [10.129.2.45/23] from ovn-kubernetes logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:16 +0000 UTC Normal Pod check-span-55pn5.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:16 +0000 UTC Normal Pod check-span-55pn5.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:16 +0000 UTC Normal Pod check-span-55pn5.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:16 +0000 UTC Warning Pod myapp-78ddfc878d-pmnwq.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.128.2.73:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:16 +0000 UTC Normal Pod report-span-v7sgp AddedInterface Add eth0 [10.129.2.44/23] from ovn-kubernetes logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:16 +0000 UTC Normal Pod report-span-v7sgp.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:16 +0000 UTC Normal Pod report-span-v7sgp.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:16 +0000 UTC Normal Pod report-span-v7sgp.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | 2024-01-08 07:31:26 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:31:27 | examples-business-application-injected-sidecar | Deleting namespace: kuttl-test-inviting-zebra === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (1279.84s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.44s) --- PASS: kuttl/harness/examples-simple-prod (104.73s) --- PASS: kuttl/harness/examples-with-sampling (60.60s) --- PASS: kuttl/harness/examples-with-cassandra (59.03s) --- PASS: kuttl/harness/examples-with-badger-and-volume (41.15s) --- PASS: kuttl/harness/examples-with-badger (41.95s) --- PASS: kuttl/harness/examples-simplest (41.08s) --- PASS: kuttl/harness/examples-simple-prod-with-volumes (74.54s) --- PASS: kuttl/harness/examples-openshift-with-htpasswd (24.42s) --- PASS: kuttl/harness/examples-service-types (53.88s) --- FAIL: kuttl/harness/examples-collector-with-priority-class (628.98s) --- PASS: kuttl/harness/examples-agent-with-priority-class (40.26s) --- PASS: kuttl/harness/examples-all-in-one-with-options (40.76s) --- PASS: kuttl/harness/examples-agent-as-daemonset (17.57s) --- PASS: kuttl/harness/examples-business-application-injected-sidecar (44.27s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name examples --report --output /logs/artifacts/examples.xml ./artifacts/kuttl-report.xml time="2024-01-08T07:31:42Z" level=debug msg="Setting a new name for the test suites" time="2024-01-08T07:31:42Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-01-08T07:31:42Z" level=debug msg="normalizing test case names" time="2024-01-08T07:31:42Z" level=debug msg="examples/artifacts -> examples_artifacts" time="2024-01-08T07:31:42Z" level=debug msg="examples/examples-simple-prod -> examples_examples_simple_prod" time="2024-01-08T07:31:42Z" level=debug msg="examples/examples-with-sampling -> examples_examples_with_sampling" time="2024-01-08T07:31:42Z" level=debug msg="examples/examples-with-cassandra -> examples_examples_with_cassandra" time="2024-01-08T07:31:42Z" level=debug msg="examples/examples-with-badger-and-volume -> examples_examples_with_badger_and_volume" time="2024-01-08T07:31:42Z" level=debug msg="examples/examples-with-badger -> examples_examples_with_badger" time="2024-01-08T07:31:42Z" level=debug msg="examples/examples-simplest -> examples_examples_simplest" time="2024-01-08T07:31:42Z" level=debug msg="examples/examples-simple-prod-with-volumes -> examples_examples_simple_prod_with_volumes" time="2024-01-08T07:31:42Z" level=debug msg="examples/examples-openshift-with-htpasswd -> examples_examples_openshift_with_htpasswd" time="2024-01-08T07:31:42Z" level=debug msg="examples/examples-service-types -> examples_examples_service_types" time="2024-01-08T07:31:42Z" level=debug msg="examples/examples-collector-with-priority-class -> examples_examples_collector_with_priority_class" time="2024-01-08T07:31:42Z" level=debug msg="examples/examples-agent-with-priority-class -> examples_examples_agent_with_priority_class" time="2024-01-08T07:31:42Z" level=debug msg="examples/examples-all-in-one-with-options -> examples_examples_all_in_one_with_options" time="2024-01-08T07:31:42Z" level=debug msg="examples/examples-agent-as-daemonset -> examples_examples_agent_as_daemonset" time="2024-01-08T07:31:42Z" level=debug msg="examples/examples-business-application-injected-sidecar -> examples_examples_business_application_injected_sidecar" +---------------------------------------------------------+--------+ | NAME | RESULT | +---------------------------------------------------------+--------+ | examples_artifacts | passed | | examples_examples_simple_prod | passed | | examples_examples_with_sampling | passed | | examples_examples_with_cassandra | passed | | examples_examples_with_badger_and_volume | passed | | examples_examples_with_badger | passed | | examples_examples_simplest | passed | | examples_examples_simple_prod_with_volumes | passed | | examples_examples_openshift_with_htpasswd | passed | | examples_examples_service_types | passed | | examples_examples_collector_with_priority_class | failed | | examples_examples_agent_with_priority_class | passed | | examples_examples_all_in_one_with_options | passed | | examples_examples_agent_as_daemonset | passed | | examples_examples_business_application_injected_sidecar | passed | +---------------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh generate false true + '[' 3 -ne 3 ']' + test_suite_name=generate + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/generate.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-generate make[2]: Entering directory '/tmp/jaeger-tests' test -s /tmp/jaeger-tests/bin/operator-sdk || curl -sLo /tmp/jaeger-tests/bin/operator-sdk https://github.com/operator-framework/operator-sdk/releases/download/v1.32.0/operator-sdk_`go env GOOS`_`go env GOARCH` ./hack/install/install-golangci-lint.sh Installing golangci-lint golangci-lint 1.55.2 is installed already ./hack/install/install-goimports.sh Installing goimports Try 0... go install golang.org/x/tools/cmd/goimports@v0.1.12 >>>> Formatting code... ./.ci/format.sh >>>> Building... ./hack/install/install-dependencies.sh Installing go dependencies Try 0... go mod download GOOS= GOARCH= CGO_ENABLED=0 GO111MODULE=on go build -ldflags "-X "github.com/jaegertracing/jaeger-operator/pkg/version".version="1.52.0" -X "github.com/jaegertracing/jaeger-operator/pkg/version".buildDate=2024-01-08T07:31:46Z -X "github.com/jaegertracing/jaeger-operator/pkg/version".defaultJaeger="1.52.0"" -o "bin/jaeger-operator" main.go JAEGER_VERSION="1.52.0" ./tests/e2e/generate/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-06-062415 True False 41m Cluster version is 4.15.0-0.nightly-2024-01-06-062415' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-06-062415 True False 41m Cluster version is 4.15.0-0.nightly-2024-01-06-062415' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/generate/render.sh ++ export SUITE_DIR=./tests/e2e/generate ++ SUITE_DIR=./tests/e2e/generate ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/generate ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test generate 'This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 2 -ne 2 ']' + test_name=generate + message='This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/generate/_build + '[' _build '!=' _build ']' + rm -rf generate + warning 'generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed\e[0m' WAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running generate E2E tests' Running generate E2E tests + cd tests/e2e/generate/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3521674980 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:32:39 | artifacts | Creating namespace: kuttl-test-proud-narwhal logger.go:42: 07:32:39 | artifacts | artifacts events from ns kuttl-test-proud-narwhal: logger.go:42: 07:32:39 | artifacts | Deleting namespace: kuttl-test-proud-narwhal === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.71s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.53s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name generate --report --output /logs/artifacts/generate.xml ./artifacts/kuttl-report.xml time="2024-01-08T07:32:46Z" level=debug msg="Setting a new name for the test suites" time="2024-01-08T07:32:46Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-01-08T07:32:46Z" level=debug msg="normalizing test case names" time="2024-01-08T07:32:46Z" level=debug msg="generate/artifacts -> generate_artifacts" +--------------------+--------+ | NAME | RESULT | +--------------------+--------+ | generate_artifacts | passed | +--------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh miscellaneous false true + '[' 3 -ne 3 ']' + test_suite_name=miscellaneous + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/miscellaneous.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-miscellaneous make[2]: Entering directory '/tmp/jaeger-tests' SKIP_ES_EXTERNAL=true ./tests/e2e/miscellaneous/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-06-062415 True False 41m Cluster version is 4.15.0-0.nightly-2024-01-06-062415' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-06-062415 True False 41m Cluster version is 4.15.0-0.nightly-2024-01-06-062415' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/miscellaneous/render.sh ++ export SUITE_DIR=./tests/e2e/miscellaneous ++ SUITE_DIR=./tests/e2e/miscellaneous ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/miscellaneous ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test cassandra-spark 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=cassandra-spark + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf cassandra-spark + warning 'cassandra-spark: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: cassandra-spark: Test not supported in OpenShift\e[0m' WAR: cassandra-spark: Test not supported in OpenShift + start_test collector-autoscale + '[' 1 -ne 1 ']' + test_name=collector-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-autoscale\e[0m' Rendering files for test collector-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p collector-autoscale + cd collector-autoscale + jaeger_name=simple-prod + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + ELASTICSEARCH_NODECOUNT=1 + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.resources.requests.memory="200m"' 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.autoscale=true 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.minReplicas=1 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.maxReplicas=2 01-install.yaml + version_lt 1.27 1.23 ++ echo 1.27 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.27 '!=' 1.27 + rm ./03-assert.yaml + generate_otlp_e2e_tests http + test_protocol=http + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-http\e[0m' Rendering files for test collector-otlp-allinone-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-autoscale + '[' collector-autoscale '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-http + cd collector-otlp-allinone-http + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger http true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-http\e[0m' Rendering files for test collector-otlp-production-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-http + '[' collector-otlp-allinone-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-http + cd collector-otlp-production-http + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger http true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + generate_otlp_e2e_tests grpc + test_protocol=grpc + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-grpc\e[0m' Rendering files for test collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-http + '[' collector-otlp-production-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-grpc + cd collector-otlp-allinone-grpc + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-grpc\e[0m' Rendering files for test collector-otlp-production-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-grpc + '[' collector-otlp-allinone-grpc '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-grpc + cd collector-otlp-production-grpc + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + '[' true = true ']' + skip_test istio 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=istio + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-grpc + '[' collector-otlp-production-grpc '!=' _build ']' + cd .. + rm -rf istio + warning 'istio: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: istio: Test not supported in OpenShift\e[0m' WAR: istio: Test not supported in OpenShift + '[' true = true ']' + skip_test outside-cluster 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=outside-cluster + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf outside-cluster + warning 'outside-cluster: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: outside-cluster: Test not supported in OpenShift\e[0m' WAR: outside-cluster: Test not supported in OpenShift + start_test set-custom-img + '[' 1 -ne 1 ']' + test_name=set-custom-img + echo =========================================================================== =========================================================================== + info 'Rendering files for test set-custom-img' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test set-custom-img\e[0m' Rendering files for test set-custom-img + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p set-custom-img + cd set-custom-img + jaeger_name=my-jaeger + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.image="test"' ./02-install.yaml + '[' true = true ']' + skip_test non-cluster-wide 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=non-cluster-wide + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/set-custom-img + '[' set-custom-img '!=' _build ']' + cd .. + rm -rf non-cluster-wide + warning 'non-cluster-wide: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: non-cluster-wide: Test not supported in OpenShift\e[0m' WAR: non-cluster-wide: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running miscellaneous E2E tests' Running miscellaneous E2E tests + cd tests/e2e/miscellaneous/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3521674980 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/collector-autoscale === PAUSE kuttl/harness/collector-autoscale === RUN kuttl/harness/collector-otlp-allinone-grpc === PAUSE kuttl/harness/collector-otlp-allinone-grpc === RUN kuttl/harness/collector-otlp-allinone-http === PAUSE kuttl/harness/collector-otlp-allinone-http === RUN kuttl/harness/collector-otlp-production-grpc === PAUSE kuttl/harness/collector-otlp-production-grpc === RUN kuttl/harness/collector-otlp-production-http === PAUSE kuttl/harness/collector-otlp-production-http === RUN kuttl/harness/set-custom-img === PAUSE kuttl/harness/set-custom-img === CONT kuttl/harness/artifacts logger.go:42: 07:32:57 | artifacts | Creating namespace: kuttl-test-decent-ghoul logger.go:42: 07:32:57 | artifacts | artifacts events from ns kuttl-test-decent-ghoul: logger.go:42: 07:32:57 | artifacts | Deleting namespace: kuttl-test-decent-ghoul === CONT kuttl/harness/set-custom-img logger.go:42: 07:33:04 | set-custom-img | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:33:04 | set-custom-img | Ignoring check-collector-img.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:33:04 | set-custom-img | Creating namespace: kuttl-test-literate-labrador logger.go:42: 07:33:04 | set-custom-img/1-install | starting test step 1-install logger.go:42: 07:33:04 | set-custom-img/1-install | Jaeger:kuttl-test-literate-labrador/my-jaeger created logger.go:42: 07:33:39 | set-custom-img/1-install | test step completed 1-install logger.go:42: 07:33:39 | set-custom-img/2-install | starting test step 2-install logger.go:42: 07:33:39 | set-custom-img/2-install | Jaeger:kuttl-test-literate-labrador/my-jaeger updated logger.go:42: 07:33:39 | set-custom-img/2-install | test step completed 2-install logger.go:42: 07:33:39 | set-custom-img/3-check-image | starting test step 3-check-image logger.go:42: 07:33:39 | set-custom-img/3-check-image | running command: [sh -c ./check-collector-img.sh] logger.go:42: 07:33:39 | set-custom-img/3-check-image | Collector image missmatch. Expected: test. Has: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9 logger.go:42: 07:33:45 | set-custom-img/3-check-image | Collector image asserted properly! logger.go:42: 07:33:45 | set-custom-img/3-check-image | test step completed 3-check-image logger.go:42: 07:33:45 | set-custom-img | set-custom-img events from ns kuttl-test-literate-labrador: logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:09 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestliteratelabradormyjaeger-1-69c88cbdff SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestliteratelabradormyjaeger-1-69c8pqpr7 replicaset-controller logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:09 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestliteratelabradormyjaeger-1-69c8pqpr7 Binding Scheduled Successfully assigned kuttl-test-literate-labrador/elasticsearch-cdm-kuttltestliteratelabradormyjaeger-1-69c8pqpr7 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:09 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestliteratelabradormyjaeger-1-69c8pqpr7 FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:09 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestliteratelabradormyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestliteratelabradormyjaeger-1-69c88cbdff to 1 deployment-controller logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestliteratelabradormyjaeger-1-69c8pqpr7 AddedInterface Add eth0 [10.129.2.46/23] from ovn-kubernetes logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestliteratelabradormyjaeger-1-69c8pqpr7.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestliteratelabradormyjaeger-1-69c8pqpr7.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestliteratelabradormyjaeger-1-69c8pqpr7.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestliteratelabradormyjaeger-1-69c8pqpr7.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestliteratelabradormyjaeger-1-69c8pqpr7.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestliteratelabradormyjaeger-1-69c8pqpr7.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:20 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestliteratelabradormyjaeger-1-69c8pqpr7.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:25 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestliteratelabradormyjaeger-1-69c8pqpr7.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:36 +0000 UTC Normal Pod my-jaeger-collector-796768bb59-9bfhh Binding Scheduled Successfully assigned kuttl-test-literate-labrador/my-jaeger-collector-796768bb59-9bfhh to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:36 +0000 UTC Normal Pod my-jaeger-collector-796768bb59-9bfhh AddedInterface Add eth0 [10.128.2.74/23] from ovn-kubernetes logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:36 +0000 UTC Normal Pod my-jaeger-collector-796768bb59-9bfhh.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:36 +0000 UTC Normal Pod my-jaeger-collector-796768bb59-9bfhh.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:36 +0000 UTC Normal Pod my-jaeger-collector-796768bb59-9bfhh.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:36 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-796768bb59 SuccessfulCreate Created pod: my-jaeger-collector-796768bb59-9bfhh replicaset-controller logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:36 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-796768bb59 to 1 deployment-controller logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:36 +0000 UTC Normal Pod my-jaeger-query-74d959bc6c-9lhzb Binding Scheduled Successfully assigned kuttl-test-literate-labrador/my-jaeger-query-74d959bc6c-9lhzb to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:36 +0000 UTC Normal Pod my-jaeger-query-74d959bc6c-9lhzb AddedInterface Add eth0 [10.131.0.61/23] from ovn-kubernetes logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:36 +0000 UTC Normal Pod my-jaeger-query-74d959bc6c-9lhzb.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:36 +0000 UTC Normal Pod my-jaeger-query-74d959bc6c-9lhzb.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:36 +0000 UTC Normal Pod my-jaeger-query-74d959bc6c-9lhzb.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:36 +0000 UTC Normal Pod my-jaeger-query-74d959bc6c-9lhzb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:36 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-74d959bc6c SuccessfulCreate Created pod: my-jaeger-query-74d959bc6c-9lhzb replicaset-controller logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:36 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-74d959bc6c to 1 deployment-controller logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:37 +0000 UTC Normal Pod my-jaeger-query-74d959bc6c-9lhzb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:37 +0000 UTC Normal Pod my-jaeger-query-74d959bc6c-9lhzb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:37 +0000 UTC Normal Pod my-jaeger-query-74d959bc6c-9lhzb.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:37 +0000 UTC Normal Pod my-jaeger-query-74d959bc6c-9lhzb.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:37 +0000 UTC Normal Pod my-jaeger-query-74d959bc6c-9lhzb.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:42 +0000 UTC Normal Pod my-jaeger-collector-796768bb59-9bfhh.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:42 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-796768bb59 SuccessfulDelete Deleted pod: my-jaeger-collector-796768bb59-9bfhh replicaset-controller logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:42 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-796768bb59 to 0 from 1 deployment-controller logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:43 +0000 UTC Normal Pod my-jaeger-collector-5cf7fc4757-7mtgq Binding Scheduled Successfully assigned kuttl-test-literate-labrador/my-jaeger-collector-5cf7fc4757-7mtgq to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:43 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5cf7fc4757 SuccessfulCreate Created pod: my-jaeger-collector-5cf7fc4757-7mtgq replicaset-controller logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:43 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-5cf7fc4757 to 1 deployment-controller logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:44 +0000 UTC Normal Pod my-jaeger-collector-5cf7fc4757-7mtgq AddedInterface Add eth0 [10.128.2.75/23] from ovn-kubernetes logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:44 +0000 UTC Normal Pod my-jaeger-collector-5cf7fc4757-7mtgq.spec.containers{jaeger-collector} Pulling Pulling image "test" kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:44 +0000 UTC Warning Pod my-jaeger-collector-5cf7fc4757-7mtgq.spec.containers{jaeger-collector} Failed Failed to pull image "test": reading manifest latest in docker.io/library/test: requested access to the resource is denied kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:44 +0000 UTC Warning Pod my-jaeger-collector-5cf7fc4757-7mtgq.spec.containers{jaeger-collector} Failed Error: ErrImagePull kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:44 +0000 UTC Normal Pod my-jaeger-collector-5cf7fc4757-7mtgq.spec.containers{jaeger-collector} BackOff Back-off pulling image "test" kubelet logger.go:42: 07:33:45 | set-custom-img | 2024-01-08 07:33:44 +0000 UTC Warning Pod my-jaeger-collector-5cf7fc4757-7mtgq.spec.containers{jaeger-collector} Failed Error: ImagePullBackOff kubelet logger.go:42: 07:33:45 | set-custom-img | Deleting namespace: kuttl-test-literate-labrador === CONT kuttl/harness/collector-otlp-allinone-http logger.go:42: 07:33:52 | collector-otlp-allinone-http | Creating namespace: kuttl-test-hip-whippet logger.go:42: 07:33:52 | collector-otlp-allinone-http/0-install | starting test step 0-install logger.go:42: 07:33:52 | collector-otlp-allinone-http/0-install | Jaeger:kuttl-test-hip-whippet/my-jaeger created logger.go:42: 07:33:59 | collector-otlp-allinone-http/0-install | test step completed 0-install logger.go:42: 07:33:59 | collector-otlp-allinone-http/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:33:59 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:34:01 | collector-otlp-allinone-http/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:34:07 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:34:08 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:34:08 | collector-otlp-allinone-http/1-smoke-test | job.batch/report-span created logger.go:42: 07:34:08 | collector-otlp-allinone-http/1-smoke-test | job.batch/check-span created logger.go:42: 07:34:21 | collector-otlp-allinone-http/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:34:21 | collector-otlp-allinone-http | collector-otlp-allinone-http events from ns kuttl-test-hip-whippet: logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:33:55 +0000 UTC Normal Pod my-jaeger-7cb77fdcbc-9fbq5 Binding Scheduled Successfully assigned kuttl-test-hip-whippet/my-jaeger-7cb77fdcbc-9fbq5 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:33:55 +0000 UTC Normal ReplicaSet.apps my-jaeger-7cb77fdcbc SuccessfulCreate Created pod: my-jaeger-7cb77fdcbc-9fbq5 replicaset-controller logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:33:55 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-7cb77fdcbc to 1 deployment-controller logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:33:56 +0000 UTC Normal Pod my-jaeger-7cb77fdcbc-9fbq5 AddedInterface Add eth0 [10.128.2.76/23] from ovn-kubernetes logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:33:56 +0000 UTC Normal Pod my-jaeger-7cb77fdcbc-9fbq5.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:33:56 +0000 UTC Normal Pod my-jaeger-7cb77fdcbc-9fbq5.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:33:56 +0000 UTC Normal Pod my-jaeger-7cb77fdcbc-9fbq5.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:33:56 +0000 UTC Normal Pod my-jaeger-7cb77fdcbc-9fbq5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:33:56 +0000 UTC Normal Pod my-jaeger-7cb77fdcbc-9fbq5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:33:56 +0000 UTC Normal Pod my-jaeger-7cb77fdcbc-9fbq5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:03 +0000 UTC Normal Pod my-jaeger-5fb587c4cc-gtvwc Binding Scheduled Successfully assigned kuttl-test-hip-whippet/my-jaeger-5fb587c4cc-gtvwc to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:03 +0000 UTC Normal ReplicaSet.apps my-jaeger-5fb587c4cc SuccessfulCreate Created pod: my-jaeger-5fb587c4cc-gtvwc replicaset-controller logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:03 +0000 UTC Normal Pod my-jaeger-7cb77fdcbc-9fbq5.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:03 +0000 UTC Normal Pod my-jaeger-7cb77fdcbc-9fbq5.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:03 +0000 UTC Normal ReplicaSet.apps my-jaeger-7cb77fdcbc SuccessfulDelete Deleted pod: my-jaeger-7cb77fdcbc-9fbq5 replicaset-controller logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:03 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-7cb77fdcbc to 0 from 1 deployment-controller logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:03 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5fb587c4cc to 1 deployment-controller logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:04 +0000 UTC Normal Pod my-jaeger-5fb587c4cc-gtvwc AddedInterface Add eth0 [10.128.2.77/23] from ovn-kubernetes logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:04 +0000 UTC Normal Pod my-jaeger-5fb587c4cc-gtvwc.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:04 +0000 UTC Normal Pod my-jaeger-5fb587c4cc-gtvwc.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:04 +0000 UTC Normal Pod my-jaeger-5fb587c4cc-gtvwc.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:04 +0000 UTC Normal Pod my-jaeger-5fb587c4cc-gtvwc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:04 +0000 UTC Normal Pod my-jaeger-5fb587c4cc-gtvwc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:04 +0000 UTC Normal Pod my-jaeger-5fb587c4cc-gtvwc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:08 +0000 UTC Normal Pod check-span-vdcl6 Binding Scheduled Successfully assigned kuttl-test-hip-whippet/check-span-vdcl6 to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:08 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-vdcl6 job-controller logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:08 +0000 UTC Normal Pod report-span-d9lv2 Binding Scheduled Successfully assigned kuttl-test-hip-whippet/report-span-d9lv2 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:08 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-d9lv2 job-controller logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:09 +0000 UTC Normal Pod check-span-vdcl6 AddedInterface Add eth0 [10.131.0.62/23] from ovn-kubernetes logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:09 +0000 UTC Normal Pod check-span-vdcl6.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:09 +0000 UTC Normal Pod check-span-vdcl6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:09 +0000 UTC Normal Pod check-span-vdcl6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:09 +0000 UTC Normal Pod report-span-d9lv2 AddedInterface Add eth0 [10.129.2.47/23] from ovn-kubernetes logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:09 +0000 UTC Normal Pod report-span-d9lv2.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:09 +0000 UTC Normal Pod report-span-d9lv2.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:09 +0000 UTC Normal Pod report-span-d9lv2.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:34:21 | collector-otlp-allinone-http | 2024-01-08 07:34:20 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:34:21 | collector-otlp-allinone-http | Deleting namespace: kuttl-test-hip-whippet === CONT kuttl/harness/collector-otlp-production-http logger.go:42: 07:34:33 | collector-otlp-production-http | Creating namespace: kuttl-test-robust-cod logger.go:42: 07:34:33 | collector-otlp-production-http/1-install | starting test step 1-install logger.go:42: 07:34:34 | collector-otlp-production-http/1-install | Jaeger:kuttl-test-robust-cod/my-jaeger created logger.go:42: 07:35:10 | collector-otlp-production-http/1-install | test step completed 1-install logger.go:42: 07:35:10 | collector-otlp-production-http/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:35:10 | collector-otlp-production-http/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:35:11 | collector-otlp-production-http/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:35:18 | collector-otlp-production-http/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:35:18 | collector-otlp-production-http/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:35:19 | collector-otlp-production-http/2-smoke-test | job.batch/report-span created logger.go:42: 07:35:19 | collector-otlp-production-http/2-smoke-test | job.batch/check-span created logger.go:42: 07:35:30 | collector-otlp-production-http/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:35:30 | collector-otlp-production-http | collector-otlp-production-http events from ns kuttl-test-robust-cod: logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:34:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrobustcodmyjaeger-1-868dc854c7-mpx92 Binding Scheduled Successfully assigned kuttl-test-robust-cod/elasticsearch-cdm-kuttltestrobustcodmyjaeger-1-868dc854c7-mpx92 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:34:40 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestrobustcodmyjaeger-1-868dc854c7-mpx92 FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:34:40 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestrobustcodmyjaeger-1-868dc854c7 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestrobustcodmyjaeger-1-868dc854c7-mpx92 replicaset-controller logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:34:40 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestrobustcodmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestrobustcodmyjaeger-1-868dc854c7 to 1 deployment-controller logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:34:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrobustcodmyjaeger-1-868dc854c7-mpx92 AddedInterface Add eth0 [10.129.2.48/23] from ovn-kubernetes logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:34:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrobustcodmyjaeger-1-868dc854c7-mpx92.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:34:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrobustcodmyjaeger-1-868dc854c7-mpx92.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:34:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrobustcodmyjaeger-1-868dc854c7-mpx92.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:34:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrobustcodmyjaeger-1-868dc854c7-mpx92.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:34:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrobustcodmyjaeger-1-868dc854c7-mpx92.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:34:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrobustcodmyjaeger-1-868dc854c7-mpx92.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:34:56 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestrobustcodmyjaeger-1-868dc854c7-mpx92.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:06 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7f86c7c88c SuccessfulCreate Created pod: my-jaeger-collector-7f86c7c88c-pkfnd replicaset-controller logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:06 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7f86c7c88c to 1 deployment-controller logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:06 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-8f79dc968 to 1 deployment-controller logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:07 +0000 UTC Normal Pod my-jaeger-collector-7f86c7c88c-pkfnd Binding Scheduled Successfully assigned kuttl-test-robust-cod/my-jaeger-collector-7f86c7c88c-pkfnd to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:07 +0000 UTC Normal Pod my-jaeger-collector-7f86c7c88c-pkfnd AddedInterface Add eth0 [10.128.2.78/23] from ovn-kubernetes logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:07 +0000 UTC Normal Pod my-jaeger-collector-7f86c7c88c-pkfnd.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:07 +0000 UTC Normal Pod my-jaeger-collector-7f86c7c88c-pkfnd.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:07 +0000 UTC Normal Pod my-jaeger-collector-7f86c7c88c-pkfnd.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:07 +0000 UTC Normal Pod my-jaeger-query-8f79dc968-plvhk Binding Scheduled Successfully assigned kuttl-test-robust-cod/my-jaeger-query-8f79dc968-plvhk to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:07 +0000 UTC Normal Pod my-jaeger-query-8f79dc968-plvhk AddedInterface Add eth0 [10.131.0.63/23] from ovn-kubernetes logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:07 +0000 UTC Normal Pod my-jaeger-query-8f79dc968-plvhk.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:07 +0000 UTC Normal Pod my-jaeger-query-8f79dc968-plvhk.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:07 +0000 UTC Normal Pod my-jaeger-query-8f79dc968-plvhk.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:07 +0000 UTC Normal Pod my-jaeger-query-8f79dc968-plvhk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:07 +0000 UTC Normal Pod my-jaeger-query-8f79dc968-plvhk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:07 +0000 UTC Normal Pod my-jaeger-query-8f79dc968-plvhk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:07 +0000 UTC Normal Pod my-jaeger-query-8f79dc968-plvhk.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:07 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-8f79dc968 SuccessfulCreate Created pod: my-jaeger-query-8f79dc968-plvhk replicaset-controller logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:08 +0000 UTC Normal Pod my-jaeger-query-8f79dc968-plvhk.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:08 +0000 UTC Normal Pod my-jaeger-query-8f79dc968-plvhk.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:14 +0000 UTC Normal Pod my-jaeger-query-8f79dc968-plvhk.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:14 +0000 UTC Normal Pod my-jaeger-query-8f79dc968-plvhk.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:14 +0000 UTC Normal Pod my-jaeger-query-8f79dc968-plvhk.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:14 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-8f79dc968 SuccessfulDelete Deleted pod: my-jaeger-query-8f79dc968-plvhk replicaset-controller logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:14 +0000 UTC Normal Pod my-jaeger-query-b8668cfcb-hrfpj Binding Scheduled Successfully assigned kuttl-test-robust-cod/my-jaeger-query-b8668cfcb-hrfpj to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:14 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-b8668cfcb SuccessfulCreate Created pod: my-jaeger-query-b8668cfcb-hrfpj replicaset-controller logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:14 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-8f79dc968 to 0 from 1 deployment-controller logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:14 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-b8668cfcb to 1 deployment-controller logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:15 +0000 UTC Normal Pod my-jaeger-query-b8668cfcb-hrfpj AddedInterface Add eth0 [10.131.0.64/23] from ovn-kubernetes logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:15 +0000 UTC Normal Pod my-jaeger-query-b8668cfcb-hrfpj.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:15 +0000 UTC Normal Pod my-jaeger-query-b8668cfcb-hrfpj.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:15 +0000 UTC Normal Pod my-jaeger-query-b8668cfcb-hrfpj.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:15 +0000 UTC Normal Pod my-jaeger-query-b8668cfcb-hrfpj.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:15 +0000 UTC Normal Pod my-jaeger-query-b8668cfcb-hrfpj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:15 +0000 UTC Normal Pod my-jaeger-query-b8668cfcb-hrfpj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:15 +0000 UTC Normal Pod my-jaeger-query-b8668cfcb-hrfpj.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:15 +0000 UTC Normal Pod my-jaeger-query-b8668cfcb-hrfpj.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:15 +0000 UTC Normal Pod my-jaeger-query-b8668cfcb-hrfpj.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:19 +0000 UTC Normal Pod check-span-fs8w2 Binding Scheduled Successfully assigned kuttl-test-robust-cod/check-span-fs8w2 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:19 +0000 UTC Normal Pod check-span-fs8w2 AddedInterface Add eth0 [10.128.2.80/23] from ovn-kubernetes logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:19 +0000 UTC Normal Pod check-span-fs8w2.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:19 +0000 UTC Normal Pod check-span-fs8w2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:19 +0000 UTC Normal Pod check-span-fs8w2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:19 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-fs8w2 job-controller logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:19 +0000 UTC Normal Pod report-span-znt7n Binding Scheduled Successfully assigned kuttl-test-robust-cod/report-span-znt7n to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:19 +0000 UTC Normal Pod report-span-znt7n AddedInterface Add eth0 [10.128.2.79/23] from ovn-kubernetes logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:19 +0000 UTC Normal Pod report-span-znt7n.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:19 +0000 UTC Normal Pod report-span-znt7n.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:19 +0000 UTC Normal Pod report-span-znt7n.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:19 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-znt7n job-controller logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:22 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:22 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:22 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:35:30 | collector-otlp-production-http | 2024-01-08 07:35:29 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:35:30 | collector-otlp-production-http | Deleting namespace: kuttl-test-robust-cod === CONT kuttl/harness/collector-otlp-production-grpc logger.go:42: 07:35:43 | collector-otlp-production-grpc | Creating namespace: kuttl-test-faithful-dolphin logger.go:42: 07:35:43 | collector-otlp-production-grpc/1-install | starting test step 1-install logger.go:42: 07:35:43 | collector-otlp-production-grpc/1-install | Jaeger:kuttl-test-faithful-dolphin/my-jaeger created logger.go:42: 07:36:19 | collector-otlp-production-grpc/1-install | test step completed 1-install logger.go:42: 07:36:19 | collector-otlp-production-grpc/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:36:19 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:36:21 | collector-otlp-production-grpc/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:36:27 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:36:27 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:36:28 | collector-otlp-production-grpc/2-smoke-test | job.batch/report-span created logger.go:42: 07:36:28 | collector-otlp-production-grpc/2-smoke-test | job.batch/check-span created logger.go:42: 07:36:47 | collector-otlp-production-grpc/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:36:47 | collector-otlp-production-grpc | collector-otlp-production-grpc events from ns kuttl-test-faithful-dolphin: logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:35:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfuldolphinmyjaeger-1-86c9792glr Binding Scheduled Successfully assigned kuttl-test-faithful-dolphin/elasticsearch-cdm-kuttltestfaithfuldolphinmyjaeger-1-86c9792glr to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:35:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfuldolphinmyjaeger-1-86c9792glr AddedInterface Add eth0 [10.129.2.49/23] from ovn-kubernetes logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:35:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfuldolphinmyjaeger-1-86c9792glr.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:35:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfuldolphinmyjaeger-1-86c9792glr.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:35:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfuldolphinmyjaeger-1-86c9792glr.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:35:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfuldolphinmyjaeger-1-86c9792glr.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:35:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfuldolphinmyjaeger-1-86c9792glr.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:35:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfuldolphinmyjaeger-1-86c9792glr.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:35:49 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestfaithfuldolphinmyjaeger-1-86c97bb6df SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestfaithfuldolphinmyjaeger-1-86c9792glr replicaset-controller logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:35:49 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestfaithfuldolphinmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestfaithfuldolphinmyjaeger-1-86c97bb6df to 1 deployment-controller logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:35:59 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfaithfuldolphinmyjaeger-1-86c9792glr.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:04 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfaithfuldolphinmyjaeger-1-86c9792glr.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:16 +0000 UTC Normal Pod my-jaeger-collector-666b6d95cf-llzt7 Binding Scheduled Successfully assigned kuttl-test-faithful-dolphin/my-jaeger-collector-666b6d95cf-llzt7 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:16 +0000 UTC Warning Pod my-jaeger-collector-666b6d95cf-llzt7 FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:16 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-666b6d95cf SuccessfulCreate Created pod: my-jaeger-collector-666b6d95cf-llzt7 replicaset-controller logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:16 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-666b6d95cf to 1 deployment-controller logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:16 +0000 UTC Normal Pod my-jaeger-query-78764f7999-lvb88 Binding Scheduled Successfully assigned kuttl-test-faithful-dolphin/my-jaeger-query-78764f7999-lvb88 to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:16 +0000 UTC Normal Pod my-jaeger-query-78764f7999-lvb88 AddedInterface Add eth0 [10.131.0.65/23] from ovn-kubernetes logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:16 +0000 UTC Normal Pod my-jaeger-query-78764f7999-lvb88.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:16 +0000 UTC Normal Pod my-jaeger-query-78764f7999-lvb88.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:16 +0000 UTC Normal Pod my-jaeger-query-78764f7999-lvb88.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:16 +0000 UTC Normal Pod my-jaeger-query-78764f7999-lvb88.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:16 +0000 UTC Normal Pod my-jaeger-query-78764f7999-lvb88.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:16 +0000 UTC Normal Pod my-jaeger-query-78764f7999-lvb88.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:16 +0000 UTC Normal Pod my-jaeger-query-78764f7999-lvb88.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:16 +0000 UTC Normal Pod my-jaeger-query-78764f7999-lvb88.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:16 +0000 UTC Normal Pod my-jaeger-query-78764f7999-lvb88.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:16 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-78764f7999 SuccessfulCreate Created pod: my-jaeger-query-78764f7999-lvb88 replicaset-controller logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:16 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-78764f7999 to 1 deployment-controller logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:17 +0000 UTC Normal Pod my-jaeger-collector-666b6d95cf-llzt7 AddedInterface Add eth0 [10.128.2.81/23] from ovn-kubernetes logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:17 +0000 UTC Normal Pod my-jaeger-collector-666b6d95cf-llzt7.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:17 +0000 UTC Normal Pod my-jaeger-collector-666b6d95cf-llzt7.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:17 +0000 UTC Normal Pod my-jaeger-collector-666b6d95cf-llzt7.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:23 +0000 UTC Normal Pod my-jaeger-query-78764f7999-lvb88.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:23 +0000 UTC Normal Pod my-jaeger-query-78764f7999-lvb88.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:23 +0000 UTC Normal Pod my-jaeger-query-78764f7999-lvb88.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:23 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-78764f7999 SuccessfulDelete Deleted pod: my-jaeger-query-78764f7999-lvb88 replicaset-controller logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:23 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-78764f7999 to 0 from 1 deployment-controller logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:24 +0000 UTC Normal Pod my-jaeger-query-78bbfbf9f7-6wqcx Binding Scheduled Successfully assigned kuttl-test-faithful-dolphin/my-jaeger-query-78bbfbf9f7-6wqcx to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:24 +0000 UTC Normal Pod my-jaeger-query-78bbfbf9f7-6wqcx AddedInterface Add eth0 [10.131.0.66/23] from ovn-kubernetes logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:24 +0000 UTC Normal Pod my-jaeger-query-78bbfbf9f7-6wqcx.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:24 +0000 UTC Normal Pod my-jaeger-query-78bbfbf9f7-6wqcx.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:24 +0000 UTC Normal Pod my-jaeger-query-78bbfbf9f7-6wqcx.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:24 +0000 UTC Normal Pod my-jaeger-query-78bbfbf9f7-6wqcx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:24 +0000 UTC Normal Pod my-jaeger-query-78bbfbf9f7-6wqcx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:24 +0000 UTC Normal Pod my-jaeger-query-78bbfbf9f7-6wqcx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:24 +0000 UTC Normal Pod my-jaeger-query-78bbfbf9f7-6wqcx.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:24 +0000 UTC Normal Pod my-jaeger-query-78bbfbf9f7-6wqcx.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:24 +0000 UTC Normal Pod my-jaeger-query-78bbfbf9f7-6wqcx.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:24 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-78bbfbf9f7 SuccessfulCreate Created pod: my-jaeger-query-78bbfbf9f7-6wqcx replicaset-controller logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:24 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-78bbfbf9f7 to 1 deployment-controller logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:28 +0000 UTC Normal Pod check-span-nnv7d Binding Scheduled Successfully assigned kuttl-test-faithful-dolphin/check-span-nnv7d to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:28 +0000 UTC Normal Pod check-span-nnv7d AddedInterface Add eth0 [10.128.2.83/23] from ovn-kubernetes logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:28 +0000 UTC Normal Pod check-span-nnv7d.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:28 +0000 UTC Normal Pod check-span-nnv7d.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:28 +0000 UTC Normal Pod check-span-nnv7d.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:28 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-nnv7d job-controller logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:28 +0000 UTC Normal Pod report-span-vt48q Binding Scheduled Successfully assigned kuttl-test-faithful-dolphin/report-span-vt48q to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:28 +0000 UTC Normal Pod report-span-vt48q AddedInterface Add eth0 [10.128.2.82/23] from ovn-kubernetes logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:28 +0000 UTC Normal Pod report-span-vt48q.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:28 +0000 UTC Normal Pod report-span-vt48q.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:28 +0000 UTC Normal Pod report-span-vt48q.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:28 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-vt48q job-controller logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:31 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:31 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:31 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:36:47 | collector-otlp-production-grpc | 2024-01-08 07:36:47 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:36:47 | collector-otlp-production-grpc | Deleting namespace: kuttl-test-faithful-dolphin === CONT kuttl/harness/collector-otlp-allinone-grpc logger.go:42: 07:36:55 | collector-otlp-allinone-grpc | Creating namespace: kuttl-test-mutual-wallaby logger.go:42: 07:36:55 | collector-otlp-allinone-grpc/0-install | starting test step 0-install logger.go:42: 07:36:56 | collector-otlp-allinone-grpc/0-install | Jaeger:kuttl-test-mutual-wallaby/my-jaeger created logger.go:42: 07:37:01 | collector-otlp-allinone-grpc/0-install | test step completed 0-install logger.go:42: 07:37:01 | collector-otlp-allinone-grpc/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:37:01 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:37:02 | collector-otlp-allinone-grpc/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:37:09 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344 OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:37:09 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:37:09 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/report-span created logger.go:42: 07:37:09 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/check-span created logger.go:42: 07:37:29 | collector-otlp-allinone-grpc/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | collector-otlp-allinone-grpc events from ns kuttl-test-mutual-wallaby: logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:36:59 +0000 UTC Normal Pod my-jaeger-786d95c4f9-npqk6 Binding Scheduled Successfully assigned kuttl-test-mutual-wallaby/my-jaeger-786d95c4f9-npqk6 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:36:59 +0000 UTC Normal Pod my-jaeger-786d95c4f9-npqk6 AddedInterface Add eth0 [10.128.2.84/23] from ovn-kubernetes logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:36:59 +0000 UTC Normal ReplicaSet.apps my-jaeger-786d95c4f9 SuccessfulCreate Created pod: my-jaeger-786d95c4f9-npqk6 replicaset-controller logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:36:59 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-786d95c4f9 to 1 deployment-controller logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:00 +0000 UTC Normal Pod my-jaeger-786d95c4f9-npqk6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:00 +0000 UTC Normal Pod my-jaeger-786d95c4f9-npqk6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:00 +0000 UTC Normal Pod my-jaeger-786d95c4f9-npqk6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:00 +0000 UTC Normal Pod my-jaeger-786d95c4f9-npqk6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:00 +0000 UTC Normal Pod my-jaeger-786d95c4f9-npqk6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:00 +0000 UTC Normal Pod my-jaeger-786d95c4f9-npqk6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:06 +0000 UTC Normal Pod my-jaeger-786d95c4f9-npqk6.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:06 +0000 UTC Normal Pod my-jaeger-786d95c4f9-npqk6.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:06 +0000 UTC Normal ReplicaSet.apps my-jaeger-786d95c4f9 SuccessfulDelete Deleted pod: my-jaeger-786d95c4f9-npqk6 replicaset-controller logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:06 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-786d95c4f9 to 0 from 1 deployment-controller logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:07 +0000 UTC Normal Pod my-jaeger-6666fcf97c-2f66f Binding Scheduled Successfully assigned kuttl-test-mutual-wallaby/my-jaeger-6666fcf97c-2f66f to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:07 +0000 UTC Normal Pod my-jaeger-6666fcf97c-2f66f AddedInterface Add eth0 [10.131.0.67/23] from ovn-kubernetes logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:07 +0000 UTC Normal Pod my-jaeger-6666fcf97c-2f66f.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:07 +0000 UTC Normal Pod my-jaeger-6666fcf97c-2f66f.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:07 +0000 UTC Normal Pod my-jaeger-6666fcf97c-2f66f.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:07 +0000 UTC Normal Pod my-jaeger-6666fcf97c-2f66f.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:07 +0000 UTC Normal Pod my-jaeger-6666fcf97c-2f66f.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:07 +0000 UTC Normal Pod my-jaeger-6666fcf97c-2f66f.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:07 +0000 UTC Normal ReplicaSet.apps my-jaeger-6666fcf97c SuccessfulCreate Created pod: my-jaeger-6666fcf97c-2f66f replicaset-controller logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:07 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6666fcf97c to 1 deployment-controller logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:09 +0000 UTC Normal Pod check-span-5mqt6 Binding Scheduled Successfully assigned kuttl-test-mutual-wallaby/check-span-5mqt6 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:09 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-5mqt6 job-controller logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:09 +0000 UTC Normal Pod report-span-66dn5 Binding Scheduled Successfully assigned kuttl-test-mutual-wallaby/report-span-66dn5 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:09 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-66dn5 job-controller logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:10 +0000 UTC Normal Pod check-span-5mqt6 AddedInterface Add eth0 [10.128.2.85/23] from ovn-kubernetes logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:10 +0000 UTC Normal Pod check-span-5mqt6.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:10 +0000 UTC Normal Pod check-span-5mqt6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:10 +0000 UTC Normal Pod check-span-5mqt6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:10 +0000 UTC Normal Pod report-span-66dn5 AddedInterface Add eth0 [10.129.2.50/23] from ovn-kubernetes logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:10 +0000 UTC Normal Pod report-span-66dn5.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:10 +0000 UTC Normal Pod report-span-66dn5.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:10 +0000 UTC Normal Pod report-span-66dn5.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | 2024-01-08 07:37:29 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:37:29 | collector-otlp-allinone-grpc | Deleting namespace: kuttl-test-mutual-wallaby === CONT kuttl/harness/collector-autoscale logger.go:42: 07:37:42 | collector-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:37:42 | collector-autoscale | Creating namespace: kuttl-test-loyal-teal logger.go:42: 07:37:42 | collector-autoscale/1-install | starting test step 1-install logger.go:42: 07:37:42 | collector-autoscale/1-install | Jaeger:kuttl-test-loyal-teal/simple-prod created logger.go:42: 07:38:19 | collector-autoscale/1-install | test step completed 1-install logger.go:42: 07:38:19 | collector-autoscale/2- | starting test step 2- logger.go:42: 07:38:19 | collector-autoscale/2- | test step completed 2- logger.go:42: 07:38:19 | collector-autoscale | collector-autoscale events from ns kuttl-test-loyal-teal: logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:37:49 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestloyaltealsimpleprod-1-5768548785 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestloyaltealsimpleprod-1-576854878r5ntj replicaset-controller logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:37:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestloyaltealsimpleprod-1-576854878r5ntj Binding Scheduled Successfully assigned kuttl-test-loyal-teal/elasticsearch-cdm-kuttltestloyaltealsimpleprod-1-576854878r5ntj to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:37:49 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestloyaltealsimpleprod-1-576854878r5ntj FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:37:49 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestloyaltealsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestloyaltealsimpleprod-1-5768548785 to 1 deployment-controller logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:37:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestloyaltealsimpleprod-1-576854878r5ntj AddedInterface Add eth0 [10.129.2.51/23] from ovn-kubernetes logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:37:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestloyaltealsimpleprod-1-576854878r5ntj.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:37:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestloyaltealsimpleprod-1-576854878r5ntj.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:37:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestloyaltealsimpleprod-1-576854878r5ntj.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:37:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestloyaltealsimpleprod-1-576854878r5ntj.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:37:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestloyaltealsimpleprod-1-576854878r5ntj.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:37:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestloyaltealsimpleprod-1-576854878r5ntj.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:00 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestloyaltealsimpleprod-1-576854878r5ntj.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:05 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestloyaltealsimpleprod-1-576854878r5ntj.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:15 +0000 UTC Normal Pod simple-prod-collector-676b7978c-zmbzg Binding Scheduled Successfully assigned kuttl-test-loyal-teal/simple-prod-collector-676b7978c-zmbzg to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:15 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-676b7978c SuccessfulCreate Created pod: simple-prod-collector-676b7978c-zmbzg replicaset-controller logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:15 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-676b7978c to 1 deployment-controller logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:15 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5948dd7f4c SuccessfulCreate Created pod: simple-prod-query-5948dd7f4c-czms6 replicaset-controller logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:15 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-5948dd7f4c to 1 deployment-controller logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:16 +0000 UTC Normal Pod simple-prod-collector-676b7978c-zmbzg AddedInterface Add eth0 [10.128.2.86/23] from ovn-kubernetes logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:16 +0000 UTC Normal Pod simple-prod-collector-676b7978c-zmbzg.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:16 +0000 UTC Normal Pod simple-prod-collector-676b7978c-zmbzg.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:16 +0000 UTC Normal Pod simple-prod-collector-676b7978c-zmbzg.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:16 +0000 UTC Normal Pod simple-prod-query-5948dd7f4c-czms6 Binding Scheduled Successfully assigned kuttl-test-loyal-teal/simple-prod-query-5948dd7f4c-czms6 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:16 +0000 UTC Normal Pod simple-prod-query-5948dd7f4c-czms6 AddedInterface Add eth0 [10.128.2.87/23] from ovn-kubernetes logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:16 +0000 UTC Normal Pod simple-prod-query-5948dd7f4c-czms6.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:16 +0000 UTC Normal Pod simple-prod-query-5948dd7f4c-czms6.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:16 +0000 UTC Normal Pod simple-prod-query-5948dd7f4c-czms6.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:16 +0000 UTC Normal Pod simple-prod-query-5948dd7f4c-czms6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:16 +0000 UTC Normal Pod simple-prod-query-5948dd7f4c-czms6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:16 +0000 UTC Normal Pod simple-prod-query-5948dd7f4c-czms6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:16 +0000 UTC Normal Pod simple-prod-query-5948dd7f4c-czms6.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:16 +0000 UTC Normal Pod simple-prod-query-5948dd7f4c-czms6.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:38:19 | collector-autoscale | 2024-01-08 07:38:16 +0000 UTC Normal Pod simple-prod-query-5948dd7f4c-czms6.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:38:19 | collector-autoscale | Deleting namespace: kuttl-test-loyal-teal === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (328.21s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.33s) --- PASS: kuttl/harness/set-custom-img (48.11s) --- PASS: kuttl/harness/collector-otlp-allinone-http (41.67s) --- PASS: kuttl/harness/collector-otlp-production-http (69.51s) --- PASS: kuttl/harness/collector-otlp-production-grpc (72.50s) --- PASS: kuttl/harness/collector-otlp-allinone-grpc (46.69s) --- PASS: kuttl/harness/collector-autoscale (43.23s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name miscellaneous --report --output /logs/artifacts/miscellaneous.xml ./artifacts/kuttl-report.xml time="2024-01-08T07:38:26Z" level=debug msg="Setting a new name for the test suites" time="2024-01-08T07:38:26Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-01-08T07:38:26Z" level=debug msg="normalizing test case names" time="2024-01-08T07:38:26Z" level=debug msg="miscellaneous/artifacts -> miscellaneous_artifacts" time="2024-01-08T07:38:26Z" level=debug msg="miscellaneous/set-custom-img -> miscellaneous_set_custom_img" time="2024-01-08T07:38:26Z" level=debug msg="miscellaneous/collector-otlp-allinone-http -> miscellaneous_collector_otlp_allinone_http" time="2024-01-08T07:38:26Z" level=debug msg="miscellaneous/collector-otlp-production-http -> miscellaneous_collector_otlp_production_http" time="2024-01-08T07:38:26Z" level=debug msg="miscellaneous/collector-otlp-production-grpc -> miscellaneous_collector_otlp_production_grpc" time="2024-01-08T07:38:26Z" level=debug msg="miscellaneous/collector-otlp-allinone-grpc -> miscellaneous_collector_otlp_allinone_grpc" time="2024-01-08T07:38:26Z" level=debug msg="miscellaneous/collector-autoscale -> miscellaneous_collector_autoscale" +----------------------------------------------+--------+ | NAME | RESULT | +----------------------------------------------+--------+ | miscellaneous_artifacts | passed | | miscellaneous_set_custom_img | passed | | miscellaneous_collector_otlp_allinone_http | passed | | miscellaneous_collector_otlp_production_http | passed | | miscellaneous_collector_otlp_production_grpc | passed | | miscellaneous_collector_otlp_allinone_grpc | passed | | miscellaneous_collector_autoscale | passed | +----------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh sidecar false true + '[' 3 -ne 3 ']' + test_suite_name=sidecar + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/sidecar.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-sidecar make[2]: Entering directory '/tmp/jaeger-tests' ./tests/e2e/sidecar/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-06-062415 True False 46m Cluster version is 4.15.0-0.nightly-2024-01-06-062415' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-06-062415 True False 46m Cluster version is 4.15.0-0.nightly-2024-01-06-062415' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/sidecar/render.sh ++ export SUITE_DIR=./tests/e2e/sidecar ++ SUITE_DIR=./tests/e2e/sidecar ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/sidecar ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + jaeger_service_name=order + start_test sidecar-deployment + '[' 1 -ne 1 ']' + test_name=sidecar-deployment + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-deployment' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-deployment\e[0m' Rendering files for test sidecar-deployment + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build + '[' _build '!=' _build ']' + mkdir -p sidecar-deployment + cd sidecar-deployment + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-namespace + '[' 1 -ne 1 ']' + test_name=sidecar-namespace + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-namespace' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-namespace\e[0m' Rendering files for test sidecar-namespace + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-deployment + '[' sidecar-deployment '!=' _build ']' + cd .. + mkdir -p sidecar-namespace + cd sidecar-namespace + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-skip-webhook + '[' 1 -ne 1 ']' + test_name=sidecar-skip-webhook + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-skip-webhook' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-skip-webhook\e[0m' Rendering files for test sidecar-skip-webhook + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-namespace + '[' sidecar-namespace '!=' _build ']' + cd .. + mkdir -p sidecar-skip-webhook + cd sidecar-skip-webhook + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running sidecar E2E tests' Running sidecar E2E tests + cd tests/e2e/sidecar/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3521674980 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/sidecar-deployment === PAUSE kuttl/harness/sidecar-deployment === RUN kuttl/harness/sidecar-namespace === PAUSE kuttl/harness/sidecar-namespace === RUN kuttl/harness/sidecar-skip-webhook === PAUSE kuttl/harness/sidecar-skip-webhook === CONT kuttl/harness/artifacts logger.go:42: 07:38:34 | artifacts | Creating namespace: kuttl-test-game-sunfish logger.go:42: 07:38:34 | artifacts | artifacts events from ns kuttl-test-game-sunfish: logger.go:42: 07:38:34 | artifacts | Deleting namespace: kuttl-test-game-sunfish === CONT kuttl/harness/sidecar-namespace logger.go:42: 07:38:41 | sidecar-namespace | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:38:41 | sidecar-namespace | Creating namespace: kuttl-test-climbing-aardvark logger.go:42: 07:38:41 | sidecar-namespace/0-install | starting test step 0-install logger.go:42: 07:38:41 | sidecar-namespace/0-install | Jaeger:kuttl-test-climbing-aardvark/agent-as-sidecar created logger.go:42: 07:38:47 | sidecar-namespace/0-install | test step completed 0-install logger.go:42: 07:38:47 | sidecar-namespace/1-install | starting test step 1-install logger.go:42: 07:38:47 | sidecar-namespace/1-install | Deployment:kuttl-test-climbing-aardvark/vertx-create-span-sidecar created logger.go:42: 07:38:48 | sidecar-namespace/1-install | test step completed 1-install logger.go:42: 07:38:48 | sidecar-namespace/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:38:48 | sidecar-namespace/2-enable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="true"] logger.go:42: 07:38:48 | sidecar-namespace/2-enable-injection | namespace/kuttl-test-climbing-aardvark annotate logger.go:42: 07:38:53 | sidecar-namespace/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:38:53 | sidecar-namespace/3-find-service | starting test step 3-find-service logger.go:42: 07:38:53 | sidecar-namespace/3-find-service | Job:kuttl-test-climbing-aardvark/00-find-service created logger.go:42: 07:39:05 | sidecar-namespace/3-find-service | test step completed 3-find-service logger.go:42: 07:39:05 | sidecar-namespace/4-other-instance | starting test step 4-other-instance logger.go:42: 07:39:05 | sidecar-namespace/4-other-instance | Jaeger:kuttl-test-climbing-aardvark/agent-as-sidecar2 created logger.go:42: 07:39:13 | sidecar-namespace/4-other-instance | test step completed 4-other-instance logger.go:42: 07:39:13 | sidecar-namespace/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:39:13 | sidecar-namespace/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:39:13 | sidecar-namespace/6-find-service | starting test step 6-find-service logger.go:42: 07:39:13 | sidecar-namespace/6-find-service | Job:kuttl-test-climbing-aardvark/01-find-service created logger.go:42: 07:39:34 | sidecar-namespace/6-find-service | test step completed 6-find-service logger.go:42: 07:39:34 | sidecar-namespace/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:39:34 | sidecar-namespace/7-disable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="false"] logger.go:42: 07:39:34 | sidecar-namespace/7-disable-injection | namespace/kuttl-test-climbing-aardvark annotate logger.go:42: 07:39:35 | sidecar-namespace/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:39:35 | sidecar-namespace | sidecar-namespace events from ns kuttl-test-climbing-aardvark: logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:44 +0000 UTC Normal Pod agent-as-sidecar-8d5dd94dc-qpf7g Binding Scheduled Successfully assigned kuttl-test-climbing-aardvark/agent-as-sidecar-8d5dd94dc-qpf7g to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:44 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-8d5dd94dc SuccessfulCreate Created pod: agent-as-sidecar-8d5dd94dc-qpf7g replicaset-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:44 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-8d5dd94dc to 1 deployment-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:45 +0000 UTC Normal Pod agent-as-sidecar-8d5dd94dc-qpf7g AddedInterface Add eth0 [10.128.2.88/23] from ovn-kubernetes logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:45 +0000 UTC Normal Pod agent-as-sidecar-8d5dd94dc-qpf7g.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:45 +0000 UTC Normal Pod agent-as-sidecar-8d5dd94dc-qpf7g.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:45 +0000 UTC Normal Pod agent-as-sidecar-8d5dd94dc-qpf7g.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:47 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-srjb8 Binding Scheduled Successfully assigned kuttl-test-climbing-aardvark/vertx-create-span-sidecar-84d458b68c-srjb8 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:47 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-srjb8 AddedInterface Add eth0 [10.129.2.52/23] from ovn-kubernetes logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:47 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-srjb8.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:47 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-srjb8.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:47 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-srjb8.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:47 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-srjb8 replicaset-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:47 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:48 +0000 UTC Normal Pod vertx-create-span-sidecar-7fdc7b9cf-kfp2m Binding Scheduled Successfully assigned kuttl-test-climbing-aardvark/vertx-create-span-sidecar-7fdc7b9cf-kfp2m to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:48 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-7fdc7b9cf SuccessfulCreate Created pod: vertx-create-span-sidecar-7fdc7b9cf-kfp2m replicaset-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:48 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-7fdc7b9cf to 1 deployment-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:49 +0000 UTC Normal Pod vertx-create-span-sidecar-7fdc7b9cf-kfp2m AddedInterface Add eth0 [10.131.0.68/23] from ovn-kubernetes logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:49 +0000 UTC Normal Pod vertx-create-span-sidecar-7fdc7b9cf-kfp2m.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:52 +0000 UTC Normal Pod vertx-create-span-sidecar-7fdc7b9cf-kfp2m.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.682s (3.682s including waiting) kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:52 +0000 UTC Normal Pod vertx-create-span-sidecar-7fdc7b9cf-kfp2m.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:52 +0000 UTC Normal Pod vertx-create-span-sidecar-7fdc7b9cf-kfp2m.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:52 +0000 UTC Normal Pod vertx-create-span-sidecar-7fdc7b9cf-kfp2m.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:53 +0000 UTC Normal Pod 00-find-service-7gn44 Binding Scheduled Successfully assigned kuttl-test-climbing-aardvark/00-find-service-7gn44 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:53 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-7gn44 job-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:53 +0000 UTC Normal Pod vertx-create-span-sidecar-7fdc7b9cf-kfp2m.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:53 +0000 UTC Normal Pod vertx-create-span-sidecar-7fdc7b9cf-kfp2m.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:54 +0000 UTC Normal Pod 00-find-service-7gn44 AddedInterface Add eth0 [10.129.2.53/23] from ovn-kubernetes logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:54 +0000 UTC Normal Pod 00-find-service-7gn44.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:54 +0000 UTC Normal Pod 00-find-service-7gn44.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:54 +0000 UTC Normal Pod 00-find-service-7gn44.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:55 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-srjb8.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.52:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:55 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-srjb8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.52:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:58 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-srjb8.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:58 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-srjb8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.52:8080/": read tcp 10.129.2.2:48252->10.129.2.52:8080: read: connection reset by peer kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:38:58 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-srjb8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.52:8080/": dial tcp 10.129.2.52:8080: connect: connection refused kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:00 +0000 UTC Warning Pod vertx-create-span-sidecar-7fdc7b9cf-kfp2m.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.68:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:00 +0000 UTC Warning Pod vertx-create-span-sidecar-7fdc7b9cf-kfp2m.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.68:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:02 +0000 UTC Normal Pod vertx-create-span-sidecar-7fdc7b9cf-kfp2m.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:02 +0000 UTC Warning Pod vertx-create-span-sidecar-7fdc7b9cf-kfp2m.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.68:8080/": read tcp 10.131.0.2:43940->10.131.0.68:8080: read: connection reset by peer kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:02 +0000 UTC Warning Pod vertx-create-span-sidecar-7fdc7b9cf-kfp2m.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.68:8080/": dial tcp 10.131.0.68:8080: connect: connection refused kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:02 +0000 UTC Normal Pod vertx-create-span-sidecar-7fdc7b9cf-kfp2m.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:05 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:09 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-srjb8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.52:8080/": read tcp 10.129.2.2:43704->10.129.2.52:8080: read: connection reset by peer kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:11 +0000 UTC Normal Pod agent-as-sidecar2-6b5798dcfc-9d7j4 Binding Scheduled Successfully assigned kuttl-test-climbing-aardvark/agent-as-sidecar2-6b5798dcfc-9d7j4 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:11 +0000 UTC Normal Pod agent-as-sidecar2-6b5798dcfc-9d7j4 AddedInterface Add eth0 [10.128.2.89/23] from ovn-kubernetes logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:11 +0000 UTC Normal Pod agent-as-sidecar2-6b5798dcfc-9d7j4.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:11 +0000 UTC Normal Pod agent-as-sidecar2-6b5798dcfc-9d7j4.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:11 +0000 UTC Normal Pod agent-as-sidecar2-6b5798dcfc-9d7j4.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:11 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-6b5798dcfc SuccessfulCreate Created pod: agent-as-sidecar2-6b5798dcfc-9d7j4 replicaset-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:11 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-6b5798dcfc to 1 deployment-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:13 +0000 UTC Normal Pod 01-find-service-5bcm6 Binding Scheduled Successfully assigned kuttl-test-climbing-aardvark/01-find-service-5bcm6 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:13 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-5bcm6 job-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:13 +0000 UTC Normal Pod agent-as-sidecar-8d5dd94dc-qpf7g.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:14 +0000 UTC Normal Pod 01-find-service-5bcm6 AddedInterface Add eth0 [10.129.2.54/23] from ovn-kubernetes logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:14 +0000 UTC Normal Pod 01-find-service-5bcm6.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:14 +0000 UTC Normal Pod 01-find-service-5bcm6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:14 +0000 UTC Normal Pod 01-find-service-5bcm6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:14 +0000 UTC Normal Pod vertx-create-span-sidecar-6986dfff57-t2lfk Binding Scheduled Successfully assigned kuttl-test-climbing-aardvark/vertx-create-span-sidecar-6986dfff57-t2lfk to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:14 +0000 UTC Normal Pod vertx-create-span-sidecar-6986dfff57-t2lfk AddedInterface Add eth0 [10.128.2.90/23] from ovn-kubernetes logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:14 +0000 UTC Normal Pod vertx-create-span-sidecar-6986dfff57-t2lfk.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:14 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6986dfff57 SuccessfulCreate Created pod: vertx-create-span-sidecar-6986dfff57-t2lfk replicaset-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:14 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-srjb8 replicaset-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:14 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:14 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6986dfff57 to 1 from 0 deployment-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:15 +0000 UTC Normal Pod vertx-create-span-sidecar-6986dfff57-t2lfk.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:15 +0000 UTC Normal Pod vertx-create-span-sidecar-6986dfff57-t2lfk.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:15 +0000 UTC Normal Pod vertx-create-span-sidecar-6986dfff57-t2lfk.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:15 +0000 UTC Normal Pod vertx-create-span-sidecar-6986dfff57-t2lfk.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:15 +0000 UTC Normal Pod vertx-create-span-sidecar-6986dfff57-t2lfk.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:23 +0000 UTC Warning Pod vertx-create-span-sidecar-6986dfff57-t2lfk.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.90:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:23 +0000 UTC Warning Pod vertx-create-span-sidecar-6986dfff57-t2lfk.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.90:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:26 +0000 UTC Normal Pod vertx-create-span-sidecar-6986dfff57-t2lfk.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:26 +0000 UTC Warning Pod vertx-create-span-sidecar-6986dfff57-t2lfk.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.90:8080/": read tcp 10.128.2.2:49010->10.128.2.90:8080: read: connection reset by peer kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:26 +0000 UTC Warning Pod vertx-create-span-sidecar-6986dfff57-t2lfk.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.90:8080/": dial tcp 10.128.2.90:8080: connect: connection refused kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:33 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:34 +0000 UTC Normal Pod vertx-create-span-sidecar-5fd576557b-92ph4 Binding Scheduled Successfully assigned kuttl-test-climbing-aardvark/vertx-create-span-sidecar-5fd576557b-92ph4 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:34 +0000 UTC Normal Pod vertx-create-span-sidecar-5fd576557b-92ph4 AddedInterface Add eth0 [10.129.2.55/23] from ovn-kubernetes logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:34 +0000 UTC Normal Pod vertx-create-span-sidecar-5fd576557b-92ph4.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:34 +0000 UTC Normal Pod vertx-create-span-sidecar-5fd576557b-92ph4.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:34 +0000 UTC Normal Pod vertx-create-span-sidecar-5fd576557b-92ph4.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:34 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-5fd576557b SuccessfulCreate Created pod: vertx-create-span-sidecar-5fd576557b-92ph4 replicaset-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:34 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-7fdc7b9cf SuccessfulDelete Deleted pod: vertx-create-span-sidecar-7fdc7b9cf-kfp2m replicaset-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:34 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-7fdc7b9cf to 0 from 1 deployment-controller logger.go:42: 07:39:35 | sidecar-namespace | 2024-01-08 07:39:34 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-5fd576557b to 1 from 0 deployment-controller logger.go:42: 07:39:35 | sidecar-namespace | Deleting namespace: kuttl-test-climbing-aardvark === CONT kuttl/harness/sidecar-skip-webhook logger.go:42: 07:39:42 | sidecar-skip-webhook | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:39:42 | sidecar-skip-webhook | Creating namespace: kuttl-test-working-grackle logger.go:42: 07:39:42 | sidecar-skip-webhook/0-install | starting test step 0-install logger.go:42: 07:39:43 | sidecar-skip-webhook/0-install | Jaeger:kuttl-test-working-grackle/agent-as-sidecar created logger.go:42: 07:39:49 | sidecar-skip-webhook/0-install | test step completed 0-install logger.go:42: 07:39:49 | sidecar-skip-webhook/1-install | starting test step 1-install logger.go:42: 07:39:49 | sidecar-skip-webhook/1-install | Deployment:kuttl-test-working-grackle/vertx-create-span-sidecar created logger.go:42: 07:39:50 | sidecar-skip-webhook/1-install | test step completed 1-install logger.go:42: 07:39:50 | sidecar-skip-webhook/2-add-anotation-and-label | starting test step 2-add-anotation-and-label logger.go:42: 07:39:50 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name=jaeger-operator --namespace kuttl-test-working-grackle] logger.go:42: 07:39:50 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar labeled logger.go:42: 07:39:50 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-working-grackle] logger.go:42: 07:39:50 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar annotate logger.go:42: 07:39:50 | sidecar-skip-webhook/2-add-anotation-and-label | test step completed 2-add-anotation-and-label logger.go:42: 07:39:50 | sidecar-skip-webhook/3-remove-label | starting test step 3-remove-label logger.go:42: 07:39:50 | sidecar-skip-webhook/3-remove-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name- --namespace kuttl-test-working-grackle] logger.go:42: 07:39:50 | sidecar-skip-webhook/3-remove-label | deployment.apps/vertx-create-span-sidecar unlabeled logger.go:42: 07:39:52 | sidecar-skip-webhook/3-remove-label | test step completed 3-remove-label logger.go:42: 07:39:52 | sidecar-skip-webhook | sidecar-skip-webhook events from ns kuttl-test-working-grackle: logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:46 +0000 UTC Normal Pod agent-as-sidecar-69c6b65c58-2vgg4 Binding Scheduled Successfully assigned kuttl-test-working-grackle/agent-as-sidecar-69c6b65c58-2vgg4 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:46 +0000 UTC Normal Pod agent-as-sidecar-69c6b65c58-2vgg4 AddedInterface Add eth0 [10.128.2.91/23] from ovn-kubernetes logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:46 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-69c6b65c58 SuccessfulCreate Created pod: agent-as-sidecar-69c6b65c58-2vgg4 replicaset-controller logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:46 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-69c6b65c58 to 1 deployment-controller logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:47 +0000 UTC Normal Pod agent-as-sidecar-69c6b65c58-2vgg4.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:47 +0000 UTC Normal Pod agent-as-sidecar-69c6b65c58-2vgg4.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:47 +0000 UTC Normal Pod agent-as-sidecar-69c6b65c58-2vgg4.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:49 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-87bhl Binding Scheduled Successfully assigned kuttl-test-working-grackle/vertx-create-span-sidecar-84d458b68c-87bhl to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:49 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-87bhl AddedInterface Add eth0 [10.129.2.56/23] from ovn-kubernetes logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:49 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-87bhl.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:49 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-87bhl.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:49 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-87bhl.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:49 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-87bhl replicaset-controller logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:49 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:50 +0000 UTC Normal Pod vertx-create-span-sidecar-5fbb4c4c8d-ht889 Binding Scheduled Successfully assigned kuttl-test-working-grackle/vertx-create-span-sidecar-5fbb4c4c8d-ht889 to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:50 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-5fbb4c4c8d SuccessfulCreate Created pod: vertx-create-span-sidecar-5fbb4c4c8d-ht889 replicaset-controller logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:50 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-5fbb4c4c8d to 1 deployment-controller logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:51 +0000 UTC Normal Pod vertx-create-span-sidecar-5fbb4c4c8d-ht889 AddedInterface Add eth0 [10.131.0.69/23] from ovn-kubernetes logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:51 +0000 UTC Normal Pod vertx-create-span-sidecar-5fbb4c4c8d-ht889.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:51 +0000 UTC Normal Pod vertx-create-span-sidecar-5fbb4c4c8d-ht889.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:51 +0000 UTC Normal Pod vertx-create-span-sidecar-5fbb4c4c8d-ht889.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:51 +0000 UTC Normal Pod vertx-create-span-sidecar-5fbb4c4c8d-ht889.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:51 +0000 UTC Normal Pod vertx-create-span-sidecar-5fbb4c4c8d-ht889.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:39:52 | sidecar-skip-webhook | 2024-01-08 07:39:51 +0000 UTC Normal Pod vertx-create-span-sidecar-5fbb4c4c8d-ht889.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:39:52 | sidecar-skip-webhook | Deleting namespace: kuttl-test-working-grackle === CONT kuttl/harness/sidecar-deployment logger.go:42: 07:39:58 | sidecar-deployment | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:39:58 | sidecar-deployment | Creating namespace: kuttl-test-probable-tortoise logger.go:42: 07:39:58 | sidecar-deployment/0-install | starting test step 0-install logger.go:42: 07:39:58 | sidecar-deployment/0-install | Jaeger:kuttl-test-probable-tortoise/agent-as-sidecar created logger.go:42: 07:40:04 | sidecar-deployment/0-install | test step completed 0-install logger.go:42: 07:40:04 | sidecar-deployment/1-install | starting test step 1-install logger.go:42: 07:40:04 | sidecar-deployment/1-install | Deployment:kuttl-test-probable-tortoise/vertx-create-span-sidecar created logger.go:42: 07:40:06 | sidecar-deployment/1-install | test step completed 1-install logger.go:42: 07:40:06 | sidecar-deployment/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:40:06 | sidecar-deployment/2-enable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-probable-tortoise] logger.go:42: 07:40:06 | sidecar-deployment/2-enable-injection | deployment.apps/vertx-create-span-sidecar annotate logger.go:42: 07:40:07 | sidecar-deployment/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:40:07 | sidecar-deployment/3-find-service | starting test step 3-find-service logger.go:42: 07:40:07 | sidecar-deployment/3-find-service | Job:kuttl-test-probable-tortoise/00-find-service created logger.go:42: 07:40:18 | sidecar-deployment/3-find-service | test step completed 3-find-service logger.go:42: 07:40:18 | sidecar-deployment/4-other-instance | starting test step 4-other-instance logger.go:42: 07:40:18 | sidecar-deployment/4-other-instance | Jaeger:kuttl-test-probable-tortoise/agent-as-sidecar2 created logger.go:42: 07:40:24 | sidecar-deployment/4-other-instance | test step completed 4-other-instance logger.go:42: 07:40:24 | sidecar-deployment/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:40:24 | sidecar-deployment/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:40:24 | sidecar-deployment/6-find-service | starting test step 6-find-service logger.go:42: 07:40:24 | sidecar-deployment/6-find-service | Job:kuttl-test-probable-tortoise/01-find-service created logger.go:42: 07:40:44 | sidecar-deployment/6-find-service | test step completed 6-find-service logger.go:42: 07:40:44 | sidecar-deployment/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:40:44 | sidecar-deployment/7-disable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=false --namespace kuttl-test-probable-tortoise] logger.go:42: 07:40:44 | sidecar-deployment/7-disable-injection | deployment.apps/vertx-create-span-sidecar annotate logger.go:42: 07:40:46 | sidecar-deployment/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:40:47 | sidecar-deployment | sidecar-deployment events from ns kuttl-test-probable-tortoise: logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:02 +0000 UTC Normal Pod agent-as-sidecar-797f497c88-ngrdt Binding Scheduled Successfully assigned kuttl-test-probable-tortoise/agent-as-sidecar-797f497c88-ngrdt to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:02 +0000 UTC Normal Pod agent-as-sidecar-797f497c88-ngrdt AddedInterface Add eth0 [10.128.2.92/23] from ovn-kubernetes logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:02 +0000 UTC Normal Pod agent-as-sidecar-797f497c88-ngrdt.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:02 +0000 UTC Normal Pod agent-as-sidecar-797f497c88-ngrdt.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:02 +0000 UTC Normal Pod agent-as-sidecar-797f497c88-ngrdt.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:02 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-797f497c88 SuccessfulCreate Created pod: agent-as-sidecar-797f497c88-ngrdt replicaset-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:02 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-797f497c88 to 1 deployment-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:04 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-ghr64 Binding Scheduled Successfully assigned kuttl-test-probable-tortoise/vertx-create-span-sidecar-84d458b68c-ghr64 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:04 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-ghr64 AddedInterface Add eth0 [10.129.2.57/23] from ovn-kubernetes logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:04 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-ghr64.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:04 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-ghr64.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:04 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-ghr64.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:04 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-ghr64 replicaset-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:04 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:06 +0000 UTC Normal Pod vertx-create-span-sidecar-6bb544484b-7mbg7 Binding Scheduled Successfully assigned kuttl-test-probable-tortoise/vertx-create-span-sidecar-6bb544484b-7mbg7 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:06 +0000 UTC Normal Pod vertx-create-span-sidecar-6bb544484b-7mbg7 AddedInterface Add eth0 [10.129.2.58/23] from ovn-kubernetes logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:06 +0000 UTC Normal Pod vertx-create-span-sidecar-6bb544484b-7mbg7.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:06 +0000 UTC Normal Pod vertx-create-span-sidecar-6bb544484b-7mbg7.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:06 +0000 UTC Normal Pod vertx-create-span-sidecar-6bb544484b-7mbg7.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:06 +0000 UTC Normal Pod vertx-create-span-sidecar-6bb544484b-7mbg7.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:06 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6bb544484b SuccessfulCreate Created pod: vertx-create-span-sidecar-6bb544484b-7mbg7 replicaset-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:06 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6bb544484b to 1 deployment-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:07 +0000 UTC Normal Pod 00-find-service-tkrz6 Binding Scheduled Successfully assigned kuttl-test-probable-tortoise/00-find-service-tkrz6 to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:07 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-tkrz6 job-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:07 +0000 UTC Normal Pod vertx-create-span-sidecar-6bb544484b-7mbg7.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:07 +0000 UTC Normal Pod vertx-create-span-sidecar-6bb544484b-7mbg7.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:08 +0000 UTC Normal Pod 00-find-service-tkrz6 AddedInterface Add eth0 [10.131.0.70/23] from ovn-kubernetes logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:08 +0000 UTC Normal Pod 00-find-service-tkrz6.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:08 +0000 UTC Normal Pod 00-find-service-tkrz6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:08 +0000 UTC Normal Pod 00-find-service-tkrz6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:12 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-ghr64.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.57:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:12 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-ghr64.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.57:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:14 +0000 UTC Warning Pod vertx-create-span-sidecar-6bb544484b-7mbg7.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.58:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:14 +0000 UTC Warning Pod vertx-create-span-sidecar-6bb544484b-7mbg7.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.58:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:14 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-ghr64.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:15 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-ghr64.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.57:8080/": read tcp 10.129.2.2:46608->10.129.2.57:8080: read: connection reset by peer kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:15 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-ghr64.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.57:8080/": dial tcp 10.129.2.57:8080: connect: connection refused kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:16 +0000 UTC Normal Pod vertx-create-span-sidecar-6bb544484b-7mbg7.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:17 +0000 UTC Warning Pod vertx-create-span-sidecar-6bb544484b-7mbg7.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.58:8080/": read tcp 10.129.2.2:55498->10.129.2.58:8080: read: connection reset by peer kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:17 +0000 UTC Warning Pod vertx-create-span-sidecar-6bb544484b-7mbg7.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.58:8080/": dial tcp 10.129.2.58:8080: connect: connection refused kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:18 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:22 +0000 UTC Normal Pod agent-as-sidecar2-6b85899bcb-g5bds Binding Scheduled Successfully assigned kuttl-test-probable-tortoise/agent-as-sidecar2-6b85899bcb-g5bds to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:22 +0000 UTC Normal Pod agent-as-sidecar2-6b85899bcb-g5bds AddedInterface Add eth0 [10.131.0.71/23] from ovn-kubernetes logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:22 +0000 UTC Normal Pod agent-as-sidecar2-6b85899bcb-g5bds.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:22 +0000 UTC Normal Pod agent-as-sidecar2-6b85899bcb-g5bds.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:22 +0000 UTC Normal Pod agent-as-sidecar2-6b85899bcb-g5bds.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:22 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-6b85899bcb SuccessfulCreate Created pod: agent-as-sidecar2-6b85899bcb-g5bds replicaset-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:22 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-6b85899bcb to 1 deployment-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:24 +0000 UTC Normal Pod 01-find-service-5fc87 Binding Scheduled Successfully assigned kuttl-test-probable-tortoise/01-find-service-5fc87 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:24 +0000 UTC Normal Pod 01-find-service-5fc87 AddedInterface Add eth0 [10.128.2.93/23] from ovn-kubernetes logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:24 +0000 UTC Normal Pod 01-find-service-5fc87.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-03314qql/pipeline@sha256:5da00d8014d809cd6203b334b102dfdea1a17b6390a1361c4a6d8f50c0d96344" already present on machine kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:24 +0000 UTC Normal Pod 01-find-service-5fc87.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:24 +0000 UTC Normal Pod 01-find-service-5fc87.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:24 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-5fc87 job-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:24 +0000 UTC Normal Pod agent-as-sidecar-797f497c88-ngrdt.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:24 +0000 UTC Normal Pod vertx-create-span-sidecar-68758f8b64-f9hn9 Binding Scheduled Successfully assigned kuttl-test-probable-tortoise/vertx-create-span-sidecar-68758f8b64-f9hn9 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:24 +0000 UTC Warning Pod vertx-create-span-sidecar-68758f8b64-f9hn9 FailedMount MountVolume.SetUp failed for volume "agent-as-sidecar2-service-ca" : configmap references non-existent config key: service-ca.crt kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:24 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-68758f8b64 SuccessfulCreate Created pod: vertx-create-span-sidecar-68758f8b64-f9hn9 replicaset-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:24 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-ghr64.spec.containers{vertx-create-span-sidecar} Killing Stopping container vertx-create-span-sidecar kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:24 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-ghr64.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.57:8080/": read tcp 10.129.2.2:46668->10.129.2.57:8080: read: connection reset by peer kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:24 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-ghr64.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.57:8080/": read tcp 10.129.2.2:46666->10.129.2.57:8080: read: connection reset by peer kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:24 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-ghr64.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.57:8080/": dial tcp 10.129.2.57:8080: connect: connection refused kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:24 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-ghr64 replicaset-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:24 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:24 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-68758f8b64 to 1 from 0 deployment-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:25 +0000 UTC Normal Pod vertx-create-span-sidecar-68758f8b64-f9hn9 AddedInterface Add eth0 [10.128.2.94/23] from ovn-kubernetes logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:25 +0000 UTC Normal Pod vertx-create-span-sidecar-68758f8b64-f9hn9.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:25 +0000 UTC Normal Pod vertx-create-span-sidecar-68758f8b64-f9hn9.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:25 +0000 UTC Normal Pod vertx-create-span-sidecar-68758f8b64-f9hn9.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:25 +0000 UTC Normal Pod vertx-create-span-sidecar-68758f8b64-f9hn9.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:25 +0000 UTC Normal Pod vertx-create-span-sidecar-68758f8b64-f9hn9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:25 +0000 UTC Normal Pod vertx-create-span-sidecar-68758f8b64-f9hn9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:33 +0000 UTC Warning Pod vertx-create-span-sidecar-68758f8b64-f9hn9.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.94:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:33 +0000 UTC Warning Pod vertx-create-span-sidecar-68758f8b64-f9hn9.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.94:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:35 +0000 UTC Normal Pod vertx-create-span-sidecar-68758f8b64-f9hn9.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:36 +0000 UTC Warning Pod vertx-create-span-sidecar-68758f8b64-f9hn9.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.94:8080/": read tcp 10.128.2.2:40840->10.128.2.94:8080: read: connection reset by peer kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:36 +0000 UTC Warning Pod vertx-create-span-sidecar-68758f8b64-f9hn9.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.94:8080/": dial tcp 10.128.2.94:8080: connect: connection refused kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:43 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:44 +0000 UTC Normal Pod vertx-create-span-sidecar-575b94f447-p2q46 Binding Scheduled Successfully assigned kuttl-test-probable-tortoise/vertx-create-span-sidecar-575b94f447-p2q46 to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:44 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-575b94f447 SuccessfulCreate Created pod: vertx-create-span-sidecar-575b94f447-p2q46 replicaset-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:44 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6bb544484b SuccessfulDelete Deleted pod: vertx-create-span-sidecar-6bb544484b-7mbg7 replicaset-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:44 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-6bb544484b to 0 from 1 deployment-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:44 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-575b94f447 to 1 from 0 deployment-controller logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:45 +0000 UTC Normal Pod vertx-create-span-sidecar-575b94f447-p2q46 AddedInterface Add eth0 [10.131.0.72/23] from ovn-kubernetes logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:45 +0000 UTC Normal Pod vertx-create-span-sidecar-575b94f447-p2q46.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:45 +0000 UTC Normal Pod vertx-create-span-sidecar-575b94f447-p2q46.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:40:47 | sidecar-deployment | 2024-01-08 07:40:45 +0000 UTC Normal Pod vertx-create-span-sidecar-575b94f447-p2q46.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:40:47 | sidecar-deployment | Deleting namespace: kuttl-test-probable-tortoise === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (139.92s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.23s) --- PASS: kuttl/harness/sidecar-namespace (61.96s) --- PASS: kuttl/harness/sidecar-skip-webhook (15.77s) --- PASS: kuttl/harness/sidecar-deployment (55.78s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name sidecar --report --output /logs/artifacts/sidecar.xml ./artifacts/kuttl-report.xml time="2024-01-08T07:40:55Z" level=debug msg="Setting a new name for the test suites" time="2024-01-08T07:40:55Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-01-08T07:40:55Z" level=debug msg="normalizing test case names" time="2024-01-08T07:40:55Z" level=debug msg="sidecar/artifacts -> sidecar_artifacts" time="2024-01-08T07:40:55Z" level=debug msg="sidecar/sidecar-namespace -> sidecar_sidecar_namespace" time="2024-01-08T07:40:55Z" level=debug msg="sidecar/sidecar-skip-webhook -> sidecar_sidecar_skip_webhook" time="2024-01-08T07:40:55Z" level=debug msg="sidecar/sidecar-deployment -> sidecar_sidecar_deployment" +------------------------------+--------+ | NAME | RESULT | +------------------------------+--------+ | sidecar_artifacts | passed | | sidecar_sidecar_namespace | passed | | sidecar_sidecar_skip_webhook | passed | | sidecar_sidecar_deployment | passed | +------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh streaming false true + '[' 3 -ne 3 ']' + test_suite_name=streaming + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/streaming.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-streaming make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ SKIP_ES_EXTERNAL=true \ ./tests/e2e/streaming/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-06-062415 True False 49m Cluster version is 4.15.0-0.nightly-2024-01-06-062415' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-06-062415 True False 49m Cluster version is 4.15.0-0.nightly-2024-01-06-062415' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/streaming/render.sh ++ export SUITE_DIR=./tests/e2e/streaming ++ SUITE_DIR=./tests/e2e/streaming ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/streaming ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + '[' false = true ']' + start_test streaming-simple + '[' 1 -ne 1 ']' + test_name=streaming-simple + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-simple' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-simple\e[0m' Rendering files for test streaming-simple + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + mkdir -p streaming-simple + cd streaming-simple + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/streaming-jaeger-assert.yaml.template -o ./04-assert.yaml + render_smoke_test simple-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=simple-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + export JAEGER_NAME=simple-streaming + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-tls + '[' 1 -ne 1 ']' + test_name=streaming-with-tls + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-tls' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-tls\e[0m' Rendering files for test streaming-with-tls + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-simple + '[' streaming-simple '!=' _build ']' + cd .. + mkdir -p streaming-with-tls + cd streaming-with-tls + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + render_smoke_test tls-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=tls-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + export JAEGER_NAME=tls-streaming + JAEGER_NAME=tls-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-autoprovisioning-autoscale + '[' 1 -ne 1 ']' + test_name=streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-autoprovisioning-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-autoprovisioning-autoscale\e[0m' Rendering files for test streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-tls + '[' streaming-with-tls '!=' _build ']' + cd .. + mkdir -p streaming-with-autoprovisioning-autoscale + cd streaming-with-autoprovisioning-autoscale + '[' true = true ']' + rm ./00-install.yaml ./00-assert.yaml + render_install_elasticsearch upstream 01 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=01 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./01-assert.yaml + jaeger_name=auto-provisioned + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="20Mi"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="500m"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.autoscale=true ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.minReplicas=1 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.maxReplicas=2 ./02-install.yaml + render_assert_kafka true auto-provisioned 03 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provisioned + test_step=03 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./03-assert.yaml ++ expr 03 + 1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./04-assert.yaml ++ expr 03 + 2 + CLUSTER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./05-assert.yaml + version_lt 1.27 1.23 ++ echo 1.27 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.27 '!=' 1.27 + rm ./08-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running streaming E2E tests' Running streaming E2E tests + cd tests/e2e/streaming/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3521674980 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/streaming-simple === PAUSE kuttl/harness/streaming-simple === RUN kuttl/harness/streaming-with-autoprovisioning-autoscale === PAUSE kuttl/harness/streaming-with-autoprovisioning-autoscale === RUN kuttl/harness/streaming-with-tls === PAUSE kuttl/harness/streaming-with-tls === CONT kuttl/harness/artifacts logger.go:42: 07:41:08 | artifacts | Creating namespace: kuttl-test-new-elephant logger.go:42: 07:41:08 | artifacts | artifacts events from ns kuttl-test-new-elephant: logger.go:42: 07:41:08 | artifacts | Deleting namespace: kuttl-test-new-elephant === CONT kuttl/harness/streaming-with-autoprovisioning-autoscale logger.go:42: 07:41:14 | streaming-with-autoprovisioning-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:41:14 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:41:14 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:41:14 | streaming-with-autoprovisioning-autoscale | Creating namespace: kuttl-test-ideal-ray logger.go:42: 07:41:14 | streaming-with-autoprovisioning-autoscale/1-install | starting test step 1-install logger.go:42: 07:41:14 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc create sa deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:41:14 | streaming-with-autoprovisioning-autoscale/1-install | serviceaccount/deploy-elasticsearch created logger.go:42: 07:41:14 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc adm policy add-scc-to-user privileged -z deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:41:14 | streaming-with-autoprovisioning-autoscale/1-install | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:privileged added: "deploy-elasticsearch" logger.go:42: 07:41:14 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 6] logger.go:42: 07:41:20 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_0.yml -n $NAMESPACE] logger.go:42: 07:41:21 | streaming-with-autoprovisioning-autoscale/1-install | statefulset.apps/elasticsearch created logger.go:42: 07:41:21 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 3] logger.go:42: 07:41:24 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_1.yml -n $NAMESPACE] logger.go:42: 07:41:24 | streaming-with-autoprovisioning-autoscale/1-install | service/elasticsearch created logger.go:42: 07:41:42 | streaming-with-autoprovisioning-autoscale/1-install | test step completed 1-install logger.go:42: 07:41:42 | streaming-with-autoprovisioning-autoscale/2-install | starting test step 2-install logger.go:42: 07:41:42 | streaming-with-autoprovisioning-autoscale/2-install | Jaeger:kuttl-test-ideal-ray/auto-provisioned created logger.go:42: 07:41:42 | streaming-with-autoprovisioning-autoscale/2-install | test step completed 2-install logger.go:42: 07:41:42 | streaming-with-autoprovisioning-autoscale/3- | starting test step 3- logger.go:42: 07:42:58 | streaming-with-autoprovisioning-autoscale/3- | test step completed 3- logger.go:42: 07:42:58 | streaming-with-autoprovisioning-autoscale/4- | starting test step 4- logger.go:42: 07:43:30 | streaming-with-autoprovisioning-autoscale/4- | test step completed 4- logger.go:42: 07:43:30 | streaming-with-autoprovisioning-autoscale/5- | starting test step 5- logger.go:42: 07:44:04 | streaming-with-autoprovisioning-autoscale/5- | test step completed 5- logger.go:42: 07:44:04 | streaming-with-autoprovisioning-autoscale/6- | starting test step 6- logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale/6- | test step completed 6- logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale/7- | starting test step 7- logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale/7- | test step completed 7- logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | streaming-with-autoprovisioning-autoscale events from ns kuttl-test-ideal-ray: logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:41:21 +0000 UTC Normal Pod elasticsearch-0 Binding Scheduled Successfully assigned kuttl-test-ideal-ray/elasticsearch-0 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:41:21 +0000 UTC Normal Pod elasticsearch-0 AddedInterface Add eth0 [10.129.2.59/23] from ovn-kubernetes logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:41:21 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulling Pulling image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:41:21 +0000 UTC Normal StatefulSet.apps elasticsearch SuccessfulCreate create Pod elasticsearch-0 in StatefulSet elasticsearch successful statefulset-controller logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:41:29 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulled Successfully pulled image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" in 8.342s (8.342s including waiting) kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:41:29 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:41:29 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:41:36 +0000 UTC Warning Pod elasticsearch-0.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Get "http://10.129.2.59:9200/": dial tcp 10.129.2.59:9200: connect: connection refused kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:41:47 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-zookeeper NoPods No matching pods found controllermanager logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:41:47 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:41:48 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-ideal-ray/data-auto-provisioned-zookeeper-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-dc968944f-499nf_5dcb7891-0ff2-4851-b51f-c12db446648b logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:41:48 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:41:52 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ProvisioningSucceeded Successfully provisioned volume pvc-de0ee9a1-e5bb-4523-ba7d-4e2b664fcaca ebs.csi.aws.com_aws-ebs-csi-driver-controller-dc968944f-499nf_5dcb7891-0ff2-4851-b51f-c12db446648b logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:41:53 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 Binding Scheduled Successfully assigned kuttl-test-ideal-ray/auto-provisioned-zookeeper-0 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:41:54 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-de0ee9a1-e5bb-4523-ba7d-4e2b664fcaca" attachdetach-controller logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:41:58 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 AddedInterface Add eth0 [10.128.2.95/23] from ovn-kubernetes logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:41:58 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:55b18000b2e774adccd0d5393b58fc3b8358ab174eaaa3ada9d520b5cde30669" kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:42:39 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:55b18000b2e774adccd0d5393b58fc3b8358ab174eaaa3ada9d520b5cde30669" in 41.308s (41.308s including waiting) kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:42:39 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Created Created container zookeeper kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:42:39 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Started Started container zookeeper kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:42:59 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-kafka NoPods No matching pods found controllermanager logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:42:59 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:42:59 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-ideal-ray/data-0-auto-provisioned-kafka-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-dc968944f-499nf_5dcb7891-0ff2-4851-b51f-c12db446648b logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:42:59 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:03 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ProvisioningSucceeded Successfully provisioned volume pvc-741f7f5f-0ebe-4db0-bed0-b8478e52e68c ebs.csi.aws.com_aws-ebs-csi-driver-controller-dc968944f-499nf_5dcb7891-0ff2-4851-b51f-c12db446648b logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:04 +0000 UTC Normal Pod auto-provisioned-kafka-0 Binding Scheduled Successfully assigned kuttl-test-ideal-ray/auto-provisioned-kafka-0 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:06 +0000 UTC Normal Pod auto-provisioned-kafka-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-741f7f5f-0ebe-4db0-bed0-b8478e52e68c" attachdetach-controller logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:10 +0000 UTC Normal Pod auto-provisioned-kafka-0 AddedInterface Add eth0 [10.128.2.96/23] from ovn-kubernetes logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:10 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulled Container image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:55b18000b2e774adccd0d5393b58fc3b8358ab174eaaa3ada9d520b5cde30669" already present on machine kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:10 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Created Created container kafka kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:10 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Started Started container kafka kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:32 +0000 UTC Normal ReplicaSet.apps auto-provisioned-entity-operator-fc4d44bf SuccessfulCreate Created pod: auto-provisioned-entity-operator-fc4d44bf-rszn2 replicaset-controller logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:32 +0000 UTC Normal Deployment.apps auto-provisioned-entity-operator ScalingReplicaSet Scaled up replica set auto-provisioned-entity-operator-fc4d44bf to 1 deployment-controller logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:33 +0000 UTC Normal Pod auto-provisioned-entity-operator-fc4d44bf-rszn2 Binding Scheduled Successfully assigned kuttl-test-ideal-ray/auto-provisioned-entity-operator-fc4d44bf-rszn2 to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:33 +0000 UTC Normal Pod auto-provisioned-entity-operator-fc4d44bf-rszn2 AddedInterface Add eth0 [10.129.2.60/23] from ovn-kubernetes logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:33 +0000 UTC Normal Pod auto-provisioned-entity-operator-fc4d44bf-rszn2.spec.containers{topic-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:f1be1aa2f18276f9169893eb55e3733cd52fa38f2101a9b3925f79774841689f" already present on machine kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:33 +0000 UTC Normal Pod auto-provisioned-entity-operator-fc4d44bf-rszn2.spec.containers{topic-operator} Created Created container topic-operator kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:33 +0000 UTC Normal Pod auto-provisioned-entity-operator-fc4d44bf-rszn2.spec.containers{topic-operator} Started Started container topic-operator kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:33 +0000 UTC Normal Pod auto-provisioned-entity-operator-fc4d44bf-rszn2.spec.containers{user-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:f1be1aa2f18276f9169893eb55e3733cd52fa38f2101a9b3925f79774841689f" already present on machine kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:33 +0000 UTC Normal Pod auto-provisioned-entity-operator-fc4d44bf-rszn2.spec.containers{user-operator} Created Created container user-operator kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:33 +0000 UTC Normal Pod auto-provisioned-entity-operator-fc4d44bf-rszn2.spec.containers{user-operator} Started Started container user-operator kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:33 +0000 UTC Normal Pod auto-provisioned-entity-operator-fc4d44bf-rszn2.spec.containers{tls-sidecar} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:55b18000b2e774adccd0d5393b58fc3b8358ab174eaaa3ada9d520b5cde30669" kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:42 +0000 UTC Normal Pod auto-provisioned-entity-operator-fc4d44bf-rszn2.spec.containers{tls-sidecar} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:55b18000b2e774adccd0d5393b58fc3b8358ab174eaaa3ada9d520b5cde30669" in 8.682s (8.682s including waiting) kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:42 +0000 UTC Normal Pod auto-provisioned-entity-operator-fc4d44bf-rszn2.spec.containers{tls-sidecar} Created Created container tls-sidecar kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:43:42 +0000 UTC Normal Pod auto-provisioned-entity-operator-fc4d44bf-rszn2.spec.containers{tls-sidecar} Started Started container tls-sidecar kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:05 +0000 UTC Normal Pod auto-provisioned-collector-c6f654f4b-6lfkx Binding Scheduled Successfully assigned kuttl-test-ideal-ray/auto-provisioned-collector-c6f654f4b-6lfkx to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:05 +0000 UTC Normal ReplicaSet.apps auto-provisioned-collector-c6f654f4b SuccessfulCreate Created pod: auto-provisioned-collector-c6f654f4b-6lfkx replicaset-controller logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:05 +0000 UTC Normal Deployment.apps auto-provisioned-collector ScalingReplicaSet Scaled up replica set auto-provisioned-collector-c6f654f4b to 1 deployment-controller logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:05 +0000 UTC Normal Pod auto-provisioned-ingester-5cf45f57fb-ng6xv Binding Scheduled Successfully assigned kuttl-test-ideal-ray/auto-provisioned-ingester-5cf45f57fb-ng6xv to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:05 +0000 UTC Normal ReplicaSet.apps auto-provisioned-ingester-5cf45f57fb SuccessfulCreate Created pod: auto-provisioned-ingester-5cf45f57fb-ng6xv replicaset-controller logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:05 +0000 UTC Normal Deployment.apps auto-provisioned-ingester ScalingReplicaSet Scaled up replica set auto-provisioned-ingester-5cf45f57fb to 1 deployment-controller logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:05 +0000 UTC Normal Pod auto-provisioned-query-5fc7f6588c-hw2gr Binding Scheduled Successfully assigned kuttl-test-ideal-ray/auto-provisioned-query-5fc7f6588c-hw2gr to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:05 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-5fc7f6588c SuccessfulCreate Created pod: auto-provisioned-query-5fc7f6588c-hw2gr replicaset-controller logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:05 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled up replica set auto-provisioned-query-5fc7f6588c to 1 deployment-controller logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:06 +0000 UTC Normal Pod auto-provisioned-collector-c6f654f4b-6lfkx AddedInterface Add eth0 [10.131.0.73/23] from ovn-kubernetes logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:06 +0000 UTC Normal Pod auto-provisioned-collector-c6f654f4b-6lfkx.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:06 +0000 UTC Normal Pod auto-provisioned-collector-c6f654f4b-6lfkx.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:06 +0000 UTC Normal Pod auto-provisioned-collector-c6f654f4b-6lfkx.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:06 +0000 UTC Normal Pod auto-provisioned-ingester-5cf45f57fb-ng6xv AddedInterface Add eth0 [10.128.2.97/23] from ovn-kubernetes logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:06 +0000 UTC Normal Pod auto-provisioned-ingester-5cf45f57fb-ng6xv.spec.containers{jaeger-ingester} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:849018528225b7370cc4740fc9f94bef7ffd4195328a916a6013d88f885eebe2" kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:06 +0000 UTC Normal Pod auto-provisioned-query-5fc7f6588c-hw2gr AddedInterface Add eth0 [10.131.0.74/23] from ovn-kubernetes logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:06 +0000 UTC Normal Pod auto-provisioned-query-5fc7f6588c-hw2gr.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:06 +0000 UTC Normal Pod auto-provisioned-query-5fc7f6588c-hw2gr.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:06 +0000 UTC Normal Pod auto-provisioned-query-5fc7f6588c-hw2gr.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:06 +0000 UTC Normal Pod auto-provisioned-query-5fc7f6588c-hw2gr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:06 +0000 UTC Normal Pod auto-provisioned-query-5fc7f6588c-hw2gr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:06 +0000 UTC Normal Pod auto-provisioned-query-5fc7f6588c-hw2gr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:06 +0000 UTC Normal Pod auto-provisioned-query-5fc7f6588c-hw2gr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:06 +0000 UTC Normal Pod auto-provisioned-query-5fc7f6588c-hw2gr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:06 +0000 UTC Normal Pod auto-provisioned-query-5fc7f6588c-hw2gr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:20 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:20 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-ingester FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:20 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-ingester FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:21 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:21 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:21 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-ingester FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:46 +0000 UTC Normal Pod auto-provisioned-ingester-5cf45f57fb-ng6xv.spec.containers{jaeger-ingester} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:849018528225b7370cc4740fc9f94bef7ffd4195328a916a6013d88f885eebe2" in 40.075s (40.075s including waiting) kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:46 +0000 UTC Normal Pod auto-provisioned-ingester-5cf45f57fb-ng6xv.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | 2024-01-08 07:44:46 +0000 UTC Normal Pod auto-provisioned-ingester-5cf45f57fb-ng6xv.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 07:44:47 | streaming-with-autoprovisioning-autoscale | Deleting namespace: kuttl-test-ideal-ray === CONT kuttl/harness/streaming-with-tls logger.go:42: 07:45:21 | streaming-with-tls | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:45:21 | streaming-with-tls | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:45:21 | streaming-with-tls | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:45:21 | streaming-with-tls | Creating namespace: kuttl-test-wondrous-dragon logger.go:42: 07:45:21 | streaming-with-tls/0-install | starting test step 0-install logger.go:42: 07:45:21 | streaming-with-tls/0-install | running command: [sh -c cd /tmp/jaeger-tests && make undeploy-kafka KAFKA_NAMESPACE=$NAMESPACE KAFKA_OLM=true] logger.go:42: 07:45:21 | streaming-with-tls/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:45:21 | streaming-with-tls/0-install | >>>> Skiping kafka-operator undeploy logger.go:42: 07:45:21 | streaming-with-tls/0-install | kubectl delete --namespace kuttl-test-wondrous-dragon -f tests/_build/kafka-example.yaml 2>&1 || true logger.go:42: 07:45:21 | streaming-with-tls/0-install | error: the path "tests/_build/kafka-example.yaml" does not exist logger.go:42: 07:45:21 | streaming-with-tls/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:45:21 | streaming-with-tls/0-install | running command: [sh -c cd /tmp/jaeger-tests && make kafka KAFKA_NAMESPACE=$NAMESPACE KAFKA_OLM=true KAFKA_VERSION=3.6.0] logger.go:42: 07:45:21 | streaming-with-tls/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:45:21 | streaming-with-tls/0-install | >>>> Creating namespace kuttl-test-wondrous-dragon logger.go:42: 07:45:21 | streaming-with-tls/0-install | kubectl create namespace kuttl-test-wondrous-dragon 2>&1 | grep -v "already exists" || true logger.go:42: 07:45:21 | streaming-with-tls/0-install | >>>> Skipping kafka-operator deployment, assuming it has been installed via OperatorHub logger.go:42: 07:45:21 | streaming-with-tls/0-install | >>>> Creating namespace kuttl-test-wondrous-dragon logger.go:42: 07:45:21 | streaming-with-tls/0-install | mkdir -p tests/_build/ logger.go:42: 07:45:21 | streaming-with-tls/0-install | kubectl create namespace kuttl-test-wondrous-dragon 2>&1 | grep -v "already exists" || true logger.go:42: 07:45:21 | streaming-with-tls/0-install | curl --fail --location "https://raw.githubusercontent.com/strimzi/strimzi-kafka-operator/3.6.0/examples/kafka/kafka-persistent-single.yaml" --output tests/_build/kafka-example.yaml --create-dirs logger.go:42: 07:45:21 | streaming-with-tls/0-install | % Total % Received % Xferd Average Speed Time Time Time Current logger.go:42: 07:45:21 | streaming-with-tls/0-install | Dload Upload Total Spent Left Speed logger.go:42: 07:45:21 | streaming-with-tls/0-install | 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 14 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 logger.go:42: 07:45:21 | streaming-with-tls/0-install | curl: (22) The requested URL returned error: 404 logger.go:42: 07:45:21 | streaming-with-tls/0-install | make[2]: *** [Makefile:252: kafka] Error 22 logger.go:42: 07:45:21 | streaming-with-tls/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' case.go:364: failed in step 0-install case.go:366: exit status 2 logger.go:42: 07:45:21 | streaming-with-tls | streaming-with-tls events from ns kuttl-test-wondrous-dragon: logger.go:42: 07:45:21 | streaming-with-tls | Deleting namespace: kuttl-test-wondrous-dragon === CONT kuttl/harness/streaming-simple logger.go:42: 07:45:28 | streaming-simple | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:45:28 | streaming-simple | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:45:28 | streaming-simple | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:45:28 | streaming-simple | Creating namespace: kuttl-test-distinct-elephant logger.go:42: 07:45:28 | streaming-simple/0-install | starting test step 0-install logger.go:42: 07:45:28 | streaming-simple/0-install | running command: [sh -c cd /tmp/jaeger-tests && make undeploy-kafka KAFKA_NAMESPACE=$NAMESPACE KAFKA_OLM=true] logger.go:42: 07:45:28 | streaming-simple/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:45:28 | streaming-simple/0-install | >>>> Skiping kafka-operator undeploy logger.go:42: 07:45:28 | streaming-simple/0-install | kubectl delete --namespace kuttl-test-distinct-elephant -f tests/_build/kafka-example.yaml 2>&1 || true logger.go:42: 07:45:28 | streaming-simple/0-install | error: the path "tests/_build/kafka-example.yaml" does not exist logger.go:42: 07:45:28 | streaming-simple/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:45:28 | streaming-simple/0-install | running command: [sh -c cd /tmp/jaeger-tests && make kafka KAFKA_NAMESPACE=$NAMESPACE KAFKA_OLM=true KAFKA_VERSION=3.6.0] logger.go:42: 07:45:28 | streaming-simple/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:45:28 | streaming-simple/0-install | >>>> Creating namespace kuttl-test-distinct-elephant logger.go:42: 07:45:28 | streaming-simple/0-install | kubectl create namespace kuttl-test-distinct-elephant 2>&1 | grep -v "already exists" || true logger.go:42: 07:45:28 | streaming-simple/0-install | >>>> Skipping kafka-operator deployment, assuming it has been installed via OperatorHub logger.go:42: 07:45:28 | streaming-simple/0-install | >>>> Creating namespace kuttl-test-distinct-elephant logger.go:42: 07:45:28 | streaming-simple/0-install | mkdir -p tests/_build/ logger.go:42: 07:45:28 | streaming-simple/0-install | kubectl create namespace kuttl-test-distinct-elephant 2>&1 | grep -v "already exists" || true logger.go:42: 07:45:28 | streaming-simple/0-install | curl --fail --location "https://raw.githubusercontent.com/strimzi/strimzi-kafka-operator/3.6.0/examples/kafka/kafka-persistent-single.yaml" --output tests/_build/kafka-example.yaml --create-dirs logger.go:42: 07:45:28 | streaming-simple/0-install | % Total % Received % Xferd Average Speed Time Time Time Current logger.go:42: 07:45:28 | streaming-simple/0-install | Dload Upload Total Spent Left Speed logger.go:42: 07:45:28 | streaming-simple/0-install | 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 14 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 logger.go:42: 07:45:28 | streaming-simple/0-install | curl: (22) The requested URL returned error: 404 logger.go:42: 07:45:28 | streaming-simple/0-install | make[2]: *** [Makefile:252: kafka] Error 22 logger.go:42: 07:45:28 | streaming-simple/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' case.go:364: failed in step 0-install case.go:366: exit status 2 logger.go:42: 07:45:28 | streaming-simple | streaming-simple events from ns kuttl-test-distinct-elephant: logger.go:42: 07:45:28 | streaming-simple | Deleting namespace: kuttl-test-distinct-elephant === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (266.98s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.24s) --- PASS: kuttl/harness/streaming-with-autoprovisioning-autoscale (247.11s) --- FAIL: kuttl/harness/streaming-with-tls (6.80s) --- FAIL: kuttl/harness/streaming-simple (6.68s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name streaming --report --output /logs/artifacts/streaming.xml ./artifacts/kuttl-report.xml time="2024-01-08T07:45:34Z" level=debug msg="Setting a new name for the test suites" time="2024-01-08T07:45:34Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-01-08T07:45:34Z" level=debug msg="normalizing test case names" time="2024-01-08T07:45:34Z" level=debug msg="streaming/artifacts -> streaming_artifacts" time="2024-01-08T07:45:34Z" level=debug msg="streaming/streaming-with-autoprovisioning-autoscale -> streaming_streaming_with_autoprovisioning_autoscale" time="2024-01-08T07:45:34Z" level=debug msg="streaming/streaming-with-tls -> streaming_streaming_with_tls" time="2024-01-08T07:45:34Z" level=debug msg="streaming/streaming-simple -> streaming_streaming_simple" +-----------------------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------------------+--------+ | streaming_artifacts | passed | | streaming_streaming_with_autoprovisioning_autoscale | passed | | streaming_streaming_with_tls | failed | | streaming_streaming_simple | failed | +-----------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh ui false true + '[' 3 -ne 3 ']' + test_suite_name=ui + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/ui.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-ui make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true ./tests/e2e/ui/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-06-062415 True False 54m Cluster version is 4.15.0-0.nightly-2024-01-06-062415' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-06-062415 True False 54m Cluster version is 4.15.0-0.nightly-2024-01-06-062415' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/ui/render.sh ++ export SUITE_DIR=./tests/e2e/ui ++ SUITE_DIR=./tests/e2e/ui ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/ui ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test allinone + '[' 1 -ne 1 ']' + test_name=allinone + echo =========================================================================== =========================================================================== + info 'Rendering files for test allinone' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test allinone\e[0m' Rendering files for test allinone + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build + '[' _build '!=' _build ']' + mkdir -p allinone + cd allinone + export GET_URL_COMMAND + export URL + export JAEGER_NAME=all-in-one-ui + JAEGER_NAME=all-in-one-ui + '[' true = true ']' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./01-curl.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./04-test-ui-config.yaml + start_test production + '[' 1 -ne 1 ']' + test_name=production + echo =========================================================================== =========================================================================== + info 'Rendering files for test production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test production\e[0m' Rendering files for test production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build/allinone + '[' allinone '!=' _build ']' + cd .. + mkdir -p production + cd production + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + [[ true = true ]] + [[ true = true ]] + render_install_jaeger production-ui production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + '[' true = true ']' + INSECURE=true + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-forbbiden-access.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-curl.yaml + INSECURE=true + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./05-check-disabled-security.yaml + ASSERT_PRESENT=false + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./06-check-NO-gaID.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./08-check-gaID.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running ui E2E tests' Running ui E2E tests + cd tests/e2e/ui/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3521674980 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 3 tests === RUN kuttl/harness === RUN kuttl/harness/allinone === PAUSE kuttl/harness/allinone === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/production === PAUSE kuttl/harness/production === CONT kuttl/harness/allinone logger.go:42: 07:45:41 | allinone | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:45:41 | allinone | Creating namespace: kuttl-test-loving-foxhound logger.go:42: 07:45:41 | allinone/0-install | starting test step 0-install logger.go:42: 07:45:41 | allinone/0-install | Jaeger:kuttl-test-loving-foxhound/all-in-one-ui created logger.go:42: 07:45:45 | allinone/0-install | test step completed 0-install logger.go:42: 07:45:45 | allinone/1-curl | starting test step 1-curl logger.go:42: 07:45:45 | allinone/1-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 07:45:45 | allinone/1-curl | Checking the Ingress host value was populated logger.go:42: 07:45:45 | allinone/1-curl | Try number 0 logger.go:42: 07:45:45 | allinone/1-curl | Hostname is all-in-one-ui-kuttl-test-loving-foxhound.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:45:45 | allinone/1-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE all-in-one-ui] logger.go:42: 07:45:46 | allinone/1-curl | Checking an expected HTTP response logger.go:42: 07:45:46 | allinone/1-curl | Running in OpenShift logger.go:42: 07:45:46 | allinone/1-curl | User not provided. Getting the token... logger.go:42: 07:45:47 | allinone/1-curl | Warning: resource jaegers/all-in-one-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:45:53 | allinone/1-curl | Try number 1/30 the https://all-in-one-ui-kuttl-test-loving-foxhound.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:45:53 | allinone/1-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:45:53 | allinone/1-curl | Try number 2/30 the https://all-in-one-ui-kuttl-test-loving-foxhound.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:45:53 | allinone/1-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 07:46:03 | allinone/1-curl | Try number 3/30 the https://all-in-one-ui-kuttl-test-loving-foxhound.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:46:03 | allinone/1-curl | curl response asserted properly logger.go:42: 07:46:03 | allinone/1-curl | test step completed 1-curl logger.go:42: 07:46:03 | allinone/2-delete | starting test step 2-delete logger.go:42: 07:46:04 | allinone/2-delete | Jaeger:kuttl-test-loving-foxhound/all-in-one-ui created logger.go:42: 07:46:04 | allinone/2-delete | test step completed 2-delete logger.go:42: 07:46:04 | allinone/3-install | starting test step 3-install logger.go:42: 07:46:04 | allinone/3-install | Jaeger:kuttl-test-loving-foxhound/all-in-one-ui updated logger.go:42: 07:46:04 | allinone/3-install | test step completed 3-install logger.go:42: 07:46:04 | allinone/4-test-ui-config | starting test step 4-test-ui-config logger.go:42: 07:46:04 | allinone/4-test-ui-config | running command: [./ensure-ingress-host.sh] logger.go:42: 07:46:04 | allinone/4-test-ui-config | Checking the Ingress host value was populated logger.go:42: 07:46:04 | allinone/4-test-ui-config | Try number 0 logger.go:42: 07:46:04 | allinone/4-test-ui-config | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 07:46:04 | allinone/4-test-ui-config | template was: logger.go:42: 07:46:04 | allinone/4-test-ui-config | {.items[0].status.ingress[0].host} logger.go:42: 07:46:04 | allinone/4-test-ui-config | object given to jsonpath engine was: logger.go:42: 07:46:04 | allinone/4-test-ui-config | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 07:46:04 | allinone/4-test-ui-config | logger.go:42: 07:46:04 | allinone/4-test-ui-config | logger.go:42: 07:46:14 | allinone/4-test-ui-config | Try number 1 logger.go:42: 07:46:14 | allinone/4-test-ui-config | Hostname is all-in-one-ui-kuttl-test-loving-foxhound.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:46:14 | allinone/4-test-ui-config | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:46:14 | allinone/4-test-ui-config | time="2024-01-08T07:46:14Z" level=info msg="Querying https://all-in-one-ui-kuttl-test-loving-foxhound.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:46:14 | allinone/4-test-ui-config | time="2024-01-08T07:46:14Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:46:14 | allinone/4-test-ui-config | time="2024-01-08T07:46:14Z" level=info msg="Polling to https://all-in-one-ui-kuttl-test-loving-foxhound.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:46:14 | allinone/4-test-ui-config | time="2024-01-08T07:46:14Z" level=info msg="Doing request number 0" logger.go:42: 07:46:14 | allinone/4-test-ui-config | time="2024-01-08T07:46:14Z" level=warning msg="Status code: 503" logger.go:42: 07:46:22 | allinone/4-test-ui-config | time="2024-01-08T07:46:22Z" level=info msg="Doing request number 1" logger.go:42: 07:46:22 | allinone/4-test-ui-config | time="2024-01-08T07:46:22Z" level=info msg="Content found and asserted!" logger.go:42: 07:46:22 | allinone/4-test-ui-config | time="2024-01-08T07:46:22Z" level=info msg="Success!" logger.go:42: 07:46:22 | allinone/4-test-ui-config | test step completed 4-test-ui-config logger.go:42: 07:46:22 | allinone | allinone events from ns kuttl-test-loving-foxhound: logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:45 +0000 UTC Normal Pod all-in-one-ui-84885dfc87-rbp25 Binding Scheduled Successfully assigned kuttl-test-loving-foxhound/all-in-one-ui-84885dfc87-rbp25 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:45 +0000 UTC Normal Pod all-in-one-ui-84885dfc87-rbp25 AddedInterface Add eth0 [10.128.2.98/23] from ovn-kubernetes logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:45 +0000 UTC Normal Pod all-in-one-ui-84885dfc87-rbp25.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:45 +0000 UTC Normal Pod all-in-one-ui-84885dfc87-rbp25.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:45 +0000 UTC Normal Pod all-in-one-ui-84885dfc87-rbp25.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:45 +0000 UTC Normal Pod all-in-one-ui-84885dfc87-rbp25.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:45 +0000 UTC Normal Pod all-in-one-ui-84885dfc87-rbp25.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:45 +0000 UTC Normal Pod all-in-one-ui-84885dfc87-rbp25.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:45 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-84885dfc87 SuccessfulCreate Created pod: all-in-one-ui-84885dfc87-rbp25 replicaset-controller logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:45 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-84885dfc87 to 1 deployment-controller logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:47 +0000 UTC Warning Pod all-in-one-ui-84885dfc87-rbp25 FailedMount MountVolume.SetUp failed for volume "all-in-one-ui-service-ca" : configmap references non-existent config key: service-ca.crt kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:49 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-84885dfc87 SuccessfulDelete Deleted pod: all-in-one-ui-84885dfc87-rbp25 replicaset-controller logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:49 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-84885dfc87 to 0 from 1 deployment-controller logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:50 +0000 UTC Normal Pod all-in-one-ui-84885dfc87-rbp25.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:50 +0000 UTC Normal Pod all-in-one-ui-84885dfc87-rbp25.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:51 +0000 UTC Normal Pod all-in-one-ui-7bf674c4d5-vk6sk Binding Scheduled Successfully assigned kuttl-test-loving-foxhound/all-in-one-ui-7bf674c4d5-vk6sk to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:51 +0000 UTC Normal Pod all-in-one-ui-7bf674c4d5-vk6sk AddedInterface Add eth0 [10.131.0.75/23] from ovn-kubernetes logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:51 +0000 UTC Normal Pod all-in-one-ui-7bf674c4d5-vk6sk.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:51 +0000 UTC Normal Pod all-in-one-ui-7bf674c4d5-vk6sk.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:51 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-7bf674c4d5 SuccessfulCreate Created pod: all-in-one-ui-7bf674c4d5-vk6sk replicaset-controller logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:51 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-7bf674c4d5 to 1 deployment-controller logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:52 +0000 UTC Normal Pod all-in-one-ui-7bf674c4d5-vk6sk.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:52 +0000 UTC Normal Pod all-in-one-ui-7bf674c4d5-vk6sk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:52 +0000 UTC Normal Pod all-in-one-ui-7bf674c4d5-vk6sk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:45:52 +0000 UTC Normal Pod all-in-one-ui-7bf674c4d5-vk6sk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:46:04 +0000 UTC Normal Pod all-in-one-ui-7bf674c4d5-vk6sk.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:46:04 +0000 UTC Normal Pod all-in-one-ui-7bf674c4d5-vk6sk.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:46:09 +0000 UTC Normal Pod all-in-one-ui-d64b79587-b2zsx Binding Scheduled Successfully assigned kuttl-test-loving-foxhound/all-in-one-ui-d64b79587-b2zsx to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:22 | allinone | 2024-01-08 07:46:09 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-d64b79587 SuccessfulCreate Created pod: all-in-one-ui-d64b79587-b2zsx replicaset-controller logger.go:42: 07:46:22 | allinone | 2024-01-08 07:46:09 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-d64b79587 to 1 deployment-controller logger.go:42: 07:46:22 | allinone | 2024-01-08 07:46:10 +0000 UTC Normal Pod all-in-one-ui-d64b79587-b2zsx AddedInterface Add eth0 [10.128.2.99/23] from ovn-kubernetes logger.go:42: 07:46:22 | allinone | 2024-01-08 07:46:10 +0000 UTC Normal Pod all-in-one-ui-d64b79587-b2zsx.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:46:10 +0000 UTC Normal Pod all-in-one-ui-d64b79587-b2zsx.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:46:22 | allinone | 2024-01-08 07:46:10 +0000 UTC Normal Pod all-in-one-ui-d64b79587-b2zsx.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:46:23 | allinone | Deleting namespace: kuttl-test-loving-foxhound === CONT kuttl/harness/production logger.go:42: 07:46:29 | production | Ignoring add-tracking-id.yaml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:46:29 | production | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:46:29 | production | Creating namespace: kuttl-test-harmless-deer logger.go:42: 07:46:29 | production/1-install | starting test step 1-install logger.go:42: 07:46:29 | production/1-install | Jaeger:kuttl-test-harmless-deer/production-ui created logger.go:42: 07:47:04 | production/1-install | test step completed 1-install logger.go:42: 07:47:04 | production/2-check-forbbiden-access | starting test step 2-check-forbbiden-access logger.go:42: 07:47:04 | production/2-check-forbbiden-access | running command: [./ensure-ingress-host.sh] logger.go:42: 07:47:04 | production/2-check-forbbiden-access | Checking the Ingress host value was populated logger.go:42: 07:47:04 | production/2-check-forbbiden-access | Try number 0 logger.go:42: 07:47:04 | production/2-check-forbbiden-access | Hostname is production-ui-kuttl-test-harmless-deer.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:47:04 | production/2-check-forbbiden-access | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE production-ui] logger.go:42: 07:47:05 | production/2-check-forbbiden-access | Checking an expected HTTP response logger.go:42: 07:47:05 | production/2-check-forbbiden-access | Running in OpenShift logger.go:42: 07:47:05 | production/2-check-forbbiden-access | Not using any secret logger.go:42: 07:47:05 | production/2-check-forbbiden-access | Try number 1/30 the https://production-ui-kuttl-test-harmless-deer.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:05 | production/2-check-forbbiden-access | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:47:05 | production/2-check-forbbiden-access | Try number 2/30 the https://production-ui-kuttl-test-harmless-deer.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:05 | production/2-check-forbbiden-access | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 07:47:15 | production/2-check-forbbiden-access | Try number 3/30 the https://production-ui-kuttl-test-harmless-deer.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:15 | production/2-check-forbbiden-access | curl response asserted properly logger.go:42: 07:47:15 | production/2-check-forbbiden-access | test step completed 2-check-forbbiden-access logger.go:42: 07:47:15 | production/3-curl | starting test step 3-curl logger.go:42: 07:47:15 | production/3-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 07:47:15 | production/3-curl | Checking the Ingress host value was populated logger.go:42: 07:47:15 | production/3-curl | Try number 0 logger.go:42: 07:47:15 | production/3-curl | Hostname is production-ui-kuttl-test-harmless-deer.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:47:15 | production/3-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 07:47:15 | production/3-curl | Checking an expected HTTP response logger.go:42: 07:47:15 | production/3-curl | Running in OpenShift logger.go:42: 07:47:15 | production/3-curl | User not provided. Getting the token... logger.go:42: 07:47:16 | production/3-curl | Warning: resource jaegers/production-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:47:23 | production/3-curl | Try number 1/30 the https://production-ui-kuttl-test-harmless-deer.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:23 | production/3-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:47:23 | production/3-curl | Try number 2/30 the https://production-ui-kuttl-test-harmless-deer.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:23 | production/3-curl | curl response asserted properly logger.go:42: 07:47:23 | production/3-curl | test step completed 3-curl logger.go:42: 07:47:23 | production/4-install | starting test step 4-install logger.go:42: 07:47:23 | production/4-install | Jaeger:kuttl-test-harmless-deer/production-ui updated logger.go:42: 07:47:23 | production/4-install | test step completed 4-install logger.go:42: 07:47:23 | production/5-check-disabled-security | starting test step 5-check-disabled-security logger.go:42: 07:47:23 | production/5-check-disabled-security | running command: [./ensure-ingress-host.sh] logger.go:42: 07:47:23 | production/5-check-disabled-security | Checking the Ingress host value was populated logger.go:42: 07:47:23 | production/5-check-disabled-security | Try number 0 logger.go:42: 07:47:23 | production/5-check-disabled-security | Hostname is production-ui-kuttl-test-harmless-deer.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:47:23 | production/5-check-disabled-security | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 07:47:23 | production/5-check-disabled-security | Checking an expected HTTP response logger.go:42: 07:47:23 | production/5-check-disabled-security | Running in OpenShift logger.go:42: 07:47:23 | production/5-check-disabled-security | Not using any secret logger.go:42: 07:47:23 | production/5-check-disabled-security | Try number 1/30 the https://production-ui-kuttl-test-harmless-deer.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:23 | production/5-check-disabled-security | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:47:23 | production/5-check-disabled-security | Try number 2/30 the https://production-ui-kuttl-test-harmless-deer.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:23 | production/5-check-disabled-security | HTTP response is 403. 200 expected. Waiting 10 s logger.go:42: 07:47:33 | production/5-check-disabled-security | Try number 3/30 the https://production-ui-kuttl-test-harmless-deer.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:33 | production/5-check-disabled-security | curl response asserted properly logger.go:42: 07:47:33 | production/5-check-disabled-security | test step completed 5-check-disabled-security logger.go:42: 07:47:33 | production/6-check-NO-gaID | starting test step 6-check-NO-gaID logger.go:42: 07:47:33 | production/6-check-NO-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 07:47:33 | production/6-check-NO-gaID | Checking the Ingress host value was populated logger.go:42: 07:47:33 | production/6-check-NO-gaID | Try number 0 logger.go:42: 07:47:33 | production/6-check-NO-gaID | Hostname is production-ui-kuttl-test-harmless-deer.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:47:33 | production/6-check-NO-gaID | running command: [sh -c ASSERT_PRESENT=false EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:47:34 | production/6-check-NO-gaID | time="2024-01-08T07:47:34Z" level=info msg="Querying https://production-ui-kuttl-test-harmless-deer.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:47:34 | production/6-check-NO-gaID | time="2024-01-08T07:47:34Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:47:34 | production/6-check-NO-gaID | time="2024-01-08T07:47:34Z" level=info msg="Polling to https://production-ui-kuttl-test-harmless-deer.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:47:34 | production/6-check-NO-gaID | time="2024-01-08T07:47:34Z" level=info msg="Doing request number 0" logger.go:42: 07:47:34 | production/6-check-NO-gaID | time="2024-01-08T07:47:34Z" level=info msg="Content not found and asserted it was not found!" logger.go:42: 07:47:34 | production/6-check-NO-gaID | time="2024-01-08T07:47:34Z" level=info msg="Success!" logger.go:42: 07:47:34 | production/6-check-NO-gaID | test step completed 6-check-NO-gaID logger.go:42: 07:47:34 | production/7-add-tracking-id | starting test step 7-add-tracking-id logger.go:42: 07:47:34 | production/7-add-tracking-id | running command: [sh -c kubectl apply -f add-tracking-id.yaml -n $NAMESPACE] logger.go:42: 07:47:34 | production/7-add-tracking-id | jaeger.jaegertracing.io/production-ui configured logger.go:42: 07:47:34 | production/7-add-tracking-id | test step completed 7-add-tracking-id logger.go:42: 07:47:34 | production/8-check-gaID | starting test step 8-check-gaID logger.go:42: 07:47:34 | production/8-check-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 07:47:34 | production/8-check-gaID | Checking the Ingress host value was populated logger.go:42: 07:47:34 | production/8-check-gaID | Try number 0 logger.go:42: 07:47:34 | production/8-check-gaID | Hostname is production-ui-kuttl-test-harmless-deer.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:47:34 | production/8-check-gaID | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:47:35 | production/8-check-gaID | time="2024-01-08T07:47:35Z" level=info msg="Querying https://production-ui-kuttl-test-harmless-deer.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:47:35 | production/8-check-gaID | time="2024-01-08T07:47:35Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:47:35 | production/8-check-gaID | time="2024-01-08T07:47:35Z" level=info msg="Polling to https://production-ui-kuttl-test-harmless-deer.apps.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:47:35 | production/8-check-gaID | time="2024-01-08T07:47:35Z" level=info msg="Doing request number 0" logger.go:42: 07:47:35 | production/8-check-gaID | time="2024-01-08T07:47:35Z" level=warning msg="Found: false . Assert: true" logger.go:42: 07:47:35 | production/8-check-gaID | time="2024-01-08T07:47:35Z" level=warning msg="The condition of the test function was not accomplished" logger.go:42: 07:47:43 | production/8-check-gaID | time="2024-01-08T07:47:43Z" level=info msg="Doing request number 1" logger.go:42: 07:47:43 | production/8-check-gaID | time="2024-01-08T07:47:43Z" level=info msg="Content found and asserted!" logger.go:42: 07:47:43 | production/8-check-gaID | time="2024-01-08T07:47:43Z" level=info msg="Success!" logger.go:42: 07:47:43 | production/8-check-gaID | test step completed 8-check-gaID logger.go:42: 07:47:43 | production | production events from ns kuttl-test-harmless-deer: logger.go:42: 07:47:43 | production | 2024-01-08 07:46:34 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestharmlessdeerproductionui-1-788fdcfbf7 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestharmlessdeerproductionui-1-788fmpb5q replicaset-controller logger.go:42: 07:47:43 | production | 2024-01-08 07:46:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestharmlessdeerproductionui-1-788fmpb5q Binding Scheduled Successfully assigned kuttl-test-harmless-deer/elasticsearch-cdm-kuttltestharmlessdeerproductionui-1-788fmpb5q to ip-10-0-62-53.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:43 | production | 2024-01-08 07:46:34 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestharmlessdeerproductionui-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestharmlessdeerproductionui-1-788fdcfbf7 to 1 deployment-controller logger.go:42: 07:47:43 | production | 2024-01-08 07:46:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestharmlessdeerproductionui-1-788fmpb5q AddedInterface Add eth0 [10.129.2.62/23] from ovn-kubernetes logger.go:42: 07:47:43 | production | 2024-01-08 07:46:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestharmlessdeerproductionui-1-788fmpb5q.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:4587958c428b40ac31b46a96a752c2b338814895891023b3ba96ce4c12b5906d" already present on machine kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:46:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestharmlessdeerproductionui-1-788fmpb5q.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:46:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestharmlessdeerproductionui-1-788fmpb5q.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:46:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestharmlessdeerproductionui-1-788fmpb5q.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:8c14e8afb359f1b4ea4b8ed370d15d3b0c272b39ba3b90e5b314f3ccb5ac69c5" already present on machine kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:46:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestharmlessdeerproductionui-1-788fmpb5q.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:46:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestharmlessdeerproductionui-1-788fmpb5q.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:46:45 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestharmlessdeerproductionui-1-788fmpb5q.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:46:50 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestharmlessdeerproductionui-1-788fmpb5q.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:01 +0000 UTC Normal Pod production-ui-collector-7bfb69f5d7-rsh2d Binding Scheduled Successfully assigned kuttl-test-harmless-deer/production-ui-collector-7bfb69f5d7-rsh2d to ip-10-0-105-78.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:43 | production | 2024-01-08 07:47:01 +0000 UTC Normal ReplicaSet.apps production-ui-collector-7bfb69f5d7 SuccessfulCreate Created pod: production-ui-collector-7bfb69f5d7-rsh2d replicaset-controller logger.go:42: 07:47:43 | production | 2024-01-08 07:47:01 +0000 UTC Normal Deployment.apps production-ui-collector ScalingReplicaSet Scaled up replica set production-ui-collector-7bfb69f5d7 to 1 deployment-controller logger.go:42: 07:47:43 | production | 2024-01-08 07:47:01 +0000 UTC Normal Pod production-ui-query-6bcff4d955-gkrfx Binding Scheduled Successfully assigned kuttl-test-harmless-deer/production-ui-query-6bcff4d955-gkrfx to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:43 | production | 2024-01-08 07:47:01 +0000 UTC Warning Pod production-ui-query-6bcff4d955-gkrfx FailedMount MountVolume.SetUp failed for volume "production-ui-ui-oauth-proxy-tls" : secret "production-ui-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:01 +0000 UTC Normal ReplicaSet.apps production-ui-query-6bcff4d955 SuccessfulCreate Created pod: production-ui-query-6bcff4d955-gkrfx replicaset-controller logger.go:42: 07:47:43 | production | 2024-01-08 07:47:01 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-6bcff4d955 to 1 deployment-controller logger.go:42: 07:47:43 | production | 2024-01-08 07:47:02 +0000 UTC Normal Pod production-ui-collector-7bfb69f5d7-rsh2d AddedInterface Add eth0 [10.131.0.76/23] from ovn-kubernetes logger.go:42: 07:47:43 | production | 2024-01-08 07:47:02 +0000 UTC Normal Pod production-ui-collector-7bfb69f5d7-rsh2d.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:02 +0000 UTC Normal Pod production-ui-collector-7bfb69f5d7-rsh2d.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:02 +0000 UTC Normal Pod production-ui-collector-7bfb69f5d7-rsh2d.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:02 +0000 UTC Normal Pod production-ui-query-6bcff4d955-gkrfx AddedInterface Add eth0 [10.128.2.100/23] from ovn-kubernetes logger.go:42: 07:47:43 | production | 2024-01-08 07:47:02 +0000 UTC Normal Pod production-ui-query-6bcff4d955-gkrfx.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:02 +0000 UTC Normal Pod production-ui-query-6bcff4d955-gkrfx.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:02 +0000 UTC Normal Pod production-ui-query-6bcff4d955-gkrfx.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:02 +0000 UTC Normal Pod production-ui-query-6bcff4d955-gkrfx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:02 +0000 UTC Normal Pod production-ui-query-6bcff4d955-gkrfx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:02 +0000 UTC Normal Pod production-ui-query-6bcff4d955-gkrfx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:02 +0000 UTC Normal Pod production-ui-query-6bcff4d955-gkrfx.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:03 +0000 UTC Normal Pod production-ui-query-6bcff4d955-gkrfx.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:03 +0000 UTC Normal Pod production-ui-query-6bcff4d955-gkrfx.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:18 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:47:43 | production | 2024-01-08 07:47:18 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:47:43 | production | 2024-01-08 07:47:18 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:47:43 | production | 2024-01-08 07:47:18 +0000 UTC Normal Pod production-ui-query-6bcff4d955-gkrfx.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:18 +0000 UTC Normal Pod production-ui-query-6bcff4d955-gkrfx.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:18 +0000 UTC Normal Pod production-ui-query-6bcff4d955-gkrfx.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:18 +0000 UTC Normal ReplicaSet.apps production-ui-query-6bcff4d955 SuccessfulDelete Deleted pod: production-ui-query-6bcff4d955-gkrfx replicaset-controller logger.go:42: 07:47:43 | production | 2024-01-08 07:47:18 +0000 UTC Normal Pod production-ui-query-bfbcf8b87-fkx7j Binding Scheduled Successfully assigned kuttl-test-harmless-deer/production-ui-query-bfbcf8b87-fkx7j to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:43 | production | 2024-01-08 07:47:18 +0000 UTC Normal ReplicaSet.apps production-ui-query-bfbcf8b87 SuccessfulCreate Created pod: production-ui-query-bfbcf8b87-fkx7j replicaset-controller logger.go:42: 07:47:43 | production | 2024-01-08 07:47:18 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-6bcff4d955 to 0 from 1 deployment-controller logger.go:42: 07:47:43 | production | 2024-01-08 07:47:18 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-bfbcf8b87 to 1 deployment-controller logger.go:42: 07:47:43 | production | 2024-01-08 07:47:19 +0000 UTC Normal Pod production-ui-query-bfbcf8b87-fkx7j AddedInterface Add eth0 [10.128.2.101/23] from ovn-kubernetes logger.go:42: 07:47:43 | production | 2024-01-08 07:47:19 +0000 UTC Normal Pod production-ui-query-bfbcf8b87-fkx7j.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:19 +0000 UTC Normal Pod production-ui-query-bfbcf8b87-fkx7j.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:19 +0000 UTC Normal Pod production-ui-query-bfbcf8b87-fkx7j.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:19 +0000 UTC Normal Pod production-ui-query-bfbcf8b87-fkx7j.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:19 +0000 UTC Normal Pod production-ui-query-bfbcf8b87-fkx7j.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:19 +0000 UTC Normal Pod production-ui-query-bfbcf8b87-fkx7j.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:19 +0000 UTC Normal Pod production-ui-query-bfbcf8b87-fkx7j.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:19 +0000 UTC Normal Pod production-ui-query-bfbcf8b87-fkx7j.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:19 +0000 UTC Normal Pod production-ui-query-bfbcf8b87-fkx7j.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:24 +0000 UTC Normal Pod production-ui-query-bfbcf8b87-fkx7j.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:24 +0000 UTC Normal Pod production-ui-query-bfbcf8b87-fkx7j.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:24 +0000 UTC Normal Pod production-ui-query-bfbcf8b87-fkx7j.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:24 +0000 UTC Normal ReplicaSet.apps production-ui-query-bfbcf8b87 SuccessfulDelete Deleted pod: production-ui-query-bfbcf8b87-fkx7j replicaset-controller logger.go:42: 07:47:43 | production | 2024-01-08 07:47:24 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-bfbcf8b87 to 0 from 1 deployment-controller logger.go:42: 07:47:43 | production | 2024-01-08 07:47:25 +0000 UTC Normal Pod production-ui-query-7d7757b74f-db669 Binding Scheduled Successfully assigned kuttl-test-harmless-deer/production-ui-query-7d7757b74f-db669 to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:43 | production | 2024-01-08 07:47:25 +0000 UTC Normal ReplicaSet.apps production-ui-query-7d7757b74f SuccessfulCreate Created pod: production-ui-query-7d7757b74f-db669 replicaset-controller logger.go:42: 07:47:43 | production | 2024-01-08 07:47:25 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-7d7757b74f to 1 deployment-controller logger.go:42: 07:47:43 | production | 2024-01-08 07:47:26 +0000 UTC Normal Pod production-ui-query-7d7757b74f-db669 AddedInterface Add eth0 [10.128.2.102/23] from ovn-kubernetes logger.go:42: 07:47:43 | production | 2024-01-08 07:47:26 +0000 UTC Normal Pod production-ui-query-7d7757b74f-db669.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:26 +0000 UTC Normal Pod production-ui-query-7d7757b74f-db669.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:26 +0000 UTC Normal Pod production-ui-query-7d7757b74f-db669.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:26 +0000 UTC Normal Pod production-ui-query-7d7757b74f-db669.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:26 +0000 UTC Normal Pod production-ui-query-7d7757b74f-db669.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:26 +0000 UTC Normal Pod production-ui-query-7d7757b74f-db669.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:35 +0000 UTC Normal Pod production-ui-query-7d7757b74f-db669.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:35 +0000 UTC Normal Pod production-ui-query-7d7757b74f-db669.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:35 +0000 UTC Normal ReplicaSet.apps production-ui-query-7d7757b74f SuccessfulDelete Deleted pod: production-ui-query-7d7757b74f-db669 replicaset-controller logger.go:42: 07:47:43 | production | 2024-01-08 07:47:35 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-7d7757b74f to 0 from 1 deployment-controller logger.go:42: 07:47:43 | production | 2024-01-08 07:47:36 +0000 UTC Warning Pod production-ui-query-7d7757b74f-db669.spec.containers{jaeger-query} Unhealthy Readiness probe failed: Get "http://10.128.2.102:16687/": dial tcp 10.128.2.102:16687: connect: connection refused kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:36 +0000 UTC Warning Pod production-ui-query-7d7757b74f-db669.spec.containers{jaeger-agent} Unhealthy Readiness probe failed: Get "http://10.128.2.102:14271/": dial tcp 10.128.2.102:14271: connect: connection refused kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:36 +0000 UTC Normal Pod production-ui-query-d7d4cf94-ztrbx Binding Scheduled Successfully assigned kuttl-test-harmless-deer/production-ui-query-d7d4cf94-ztrbx to ip-10-0-19-211.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:43 | production | 2024-01-08 07:47:36 +0000 UTC Normal ReplicaSet.apps production-ui-query-d7d4cf94 SuccessfulCreate Created pod: production-ui-query-d7d4cf94-ztrbx replicaset-controller logger.go:42: 07:47:43 | production | 2024-01-08 07:47:36 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-d7d4cf94 to 1 deployment-controller logger.go:42: 07:47:43 | production | 2024-01-08 07:47:37 +0000 UTC Normal Pod production-ui-query-d7d4cf94-ztrbx AddedInterface Add eth0 [10.128.2.103/23] from ovn-kubernetes logger.go:42: 07:47:43 | production | 2024-01-08 07:47:37 +0000 UTC Normal Pod production-ui-query-d7d4cf94-ztrbx.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:37 +0000 UTC Normal Pod production-ui-query-d7d4cf94-ztrbx.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:37 +0000 UTC Normal Pod production-ui-query-d7d4cf94-ztrbx.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:37 +0000 UTC Normal Pod production-ui-query-d7d4cf94-ztrbx.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:37 +0000 UTC Normal Pod production-ui-query-d7d4cf94-ztrbx.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:47:43 | production | 2024-01-08 07:47:37 +0000 UTC Normal Pod production-ui-query-d7d4cf94-ztrbx.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:47:43 | production | Deleting namespace: kuttl-test-harmless-deer === CONT kuttl/harness/artifacts logger.go:42: 07:47:50 | artifacts | Creating namespace: kuttl-test-rational-marmoset logger.go:42: 07:47:50 | artifacts | artifacts events from ns kuttl-test-rational-marmoset: logger.go:42: 07:47:50 | artifacts | Deleting namespace: kuttl-test-rational-marmoset === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (135.75s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/allinone (48.10s) --- PASS: kuttl/harness/production (81.19s) --- PASS: kuttl/harness/artifacts (6.31s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name ui --report --output /logs/artifacts/ui.xml ./artifacts/kuttl-report.xml time="2024-01-08T07:47:57Z" level=debug msg="Setting a new name for the test suites" time="2024-01-08T07:47:57Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-01-08T07:47:57Z" level=debug msg="normalizing test case names" time="2024-01-08T07:47:57Z" level=debug msg="ui/allinone -> ui_allinone" time="2024-01-08T07:47:57Z" level=debug msg="ui/production -> ui_production" time="2024-01-08T07:47:57Z" level=debug msg="ui/artifacts -> ui_artifacts" +---------------+--------+ | NAME | RESULT | +---------------+--------+ | ui_allinone | passed | | ui_production | passed | | ui_artifacts | passed | +---------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh upgrade false true + '[' 3 -ne 3 ']' + test_suite_name=upgrade + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/upgrade.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-upgrade make[2]: Entering directory '/tmp/jaeger-tests' make docker JAEGER_VERSION=1.52.1 IMG="quay.io//jaeger-operator:next" make[3]: Entering directory '/tmp/jaeger-tests' [ ! -z "true" ] || docker build --build-arg=GOPROXY= --build-arg=VERSION="1.52.0" --build-arg=JAEGER_VERSION=1.52.1 --build-arg=TARGETARCH= --build-arg VERSION_DATE=2024-01-08T07:47:57Z --build-arg VERSION_PKG="github.com/jaegertracing/jaeger-operator/pkg/version" -t "quay.io//jaeger-operator:next" . make[3]: Leaving directory '/tmp/jaeger-tests' touch build-e2e-upgrade-image SKIP_ES_EXTERNAL=true IMG=quay.io//jaeger-operator:"1.52.0" JAEGER_OPERATOR_VERSION="1.52.0" JAEGER_VERSION="1.52.0" ./tests/e2e/upgrade/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-06-062415 True False 56m Cluster version is 4.15.0-0.nightly-2024-01-06-062415' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-06-062415 True False 56m Cluster version is 4.15.0-0.nightly-2024-01-06-062415' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/upgrade/render.sh ++ export SUITE_DIR=./tests/e2e/upgrade ++ SUITE_DIR=./tests/e2e/upgrade ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/upgrade ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + export JAEGER_NAME + '[' true = true ']' + skip_test upgrade 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade + warning 'upgrade: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade: Test not supported in OpenShift\e[0m' WAR: upgrade: Test not supported in OpenShift + '[' true = true ']' + skip_test upgrade-from-latest-release 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade-from-latest-release + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade-from-latest-release + warning 'upgrade-from-latest-release: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade-from-latest-release: Test not supported in OpenShift\e[0m' WAR: upgrade-from-latest-release: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running upgrade E2E tests' Running upgrade E2E tests + cd tests/e2e/upgrade/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3521674980 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-03314qql-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:47:58 | artifacts | Creating namespace: kuttl-test-legible-louse logger.go:42: 07:47:58 | artifacts | artifacts events from ns kuttl-test-legible-louse: logger.go:42: 07:47:58 | artifacts | Deleting namespace: kuttl-test-legible-louse === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.38s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.21s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name upgrade --report --output /logs/artifacts/upgrade.xml ./artifacts/kuttl-report.xml time="2024-01-08T07:48:05Z" level=debug msg="Setting a new name for the test suites" time="2024-01-08T07:48:05Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-01-08T07:48:05Z" level=debug msg="normalizing test case names" time="2024-01-08T07:48:05Z" level=debug msg="upgrade/artifacts -> upgrade_artifacts" +-------------------+--------+ | NAME | RESULT | +-------------------+--------+ | upgrade_artifacts | passed | +-------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests'