% Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 3831 100 3831 0 0 26610 0 --:--:-- --:--:-- --:--:-- 26790 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 1953 100 1953 0 0 18722 0 --:--:-- --:--:-- --:--:-- 18600 100 1953 100 1953 0 0 18713 0 --:--:-- --:--:-- --:--:-- 18600 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 350 100 350 0 0 3058 0 --:--:-- --:--:-- --:--:-- 3070 Installing kuttl Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/kubectl-kuttl https://github.com/kudobuilder/kuttl/releases/download/v0.15.0/kubectl-kuttl_0.15.0_linux_x86_64 KUBECONFIG file is: /tmp/kubeconfig-4018005699 for suite in elasticsearch examples generate miscellaneous sidecar streaming ui upgrade; do \ make run-e2e-tests-$suite ; \ done make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh elasticsearch false true + '[' 3 -ne 3 ']' + test_suite_name=elasticsearch + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/elasticsearch.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-elasticsearch make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true \ KAFKA_VERSION=0.32.0 \ SKIP_KAFKA=false \ ./tests/e2e/elasticsearch/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-11-25-110147 True False 6m28s Cluster version is 4.15.0-0.nightly-2023-11-25-110147' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-11-25-110147 True False 6m28s Cluster version is 4.15.0-0.nightly-2023-11-25-110147' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 0.32.0 ']' ++ version_le 0.32.0 0.25.0 +++ echo 0.32.0 0.25.0 +++ tr ' ' '\n' +++ head -n 1 +++ sort -V ++ test 0.25.0 == 0.32.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/elasticsearch/render.sh ++ export SUITE_DIR=./tests/e2e/elasticsearch ++ SUITE_DIR=./tests/e2e/elasticsearch ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/elasticsearch ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + start_test es-from-aio-to-production + '[' 1 -ne 1 ']' + test_name=es-from-aio-to-production + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-from-aio-to-production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-from-aio-to-production\e[0m' Rendering files for test es-from-aio-to-production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-from-aio-to-production + cd es-from-aio-to-production + jaeger_name=my-jaeger + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 03 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=03 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./03-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch.redundancyPolicy="ZeroRedundancy"' ./03-install.yaml + render_smoke_test my-jaeger true 04 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=04 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./04-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test es-increasing-replicas + '[' 1 -ne 1 ']' + test_name=es-increasing-replicas + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-increasing-replicas' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-increasing-replicas\e[0m' Rendering files for test es-increasing-replicas + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-from-aio-to-production + '[' es-from-aio-to-production '!=' _build ']' + cd .. + mkdir -p es-increasing-replicas + cd es-increasing-replicas + jaeger_name=simple-prod + '[' true = true ']' + jaeger_deployment_mode=production_autoprovisioned + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.replicas=2 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.query.replicas=2 ./02-install.yaml + cp ./01-assert.yaml ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.replicas=2 ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .status.readyReplicas=2 ./02-assert.yaml + render_smoke_test simple-prod true 03 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=03 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./03-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + cp ./02-install.yaml ./04-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.elasticsearch.nodeCount=2 ./04-install.yaml + /tmp/jaeger-tests/bin/gomplate -f ./openshift-check-es-nodes.yaml.template -o ./05-check-es-nodes.yaml + '[' true = true ']' + skip_test es-index-cleaner-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-increasing-replicas + '[' es-increasing-replicas '!=' _build ']' + cd .. + rm -rf es-index-cleaner-upstream + warning 'es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_index_cleaner -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-index-cleaner-autoprov + '[' 1 -ne 1 ']' + test_name=es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-index-cleaner-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-index-cleaner-autoprov\e[0m' Rendering files for test es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-index-cleaner-autoprov + cd es-index-cleaner-autoprov + jaeger_name=test-es-index-cleaner-with-prefix + cronjob_name=test-es-index-cleaner-with-prefix-es-index-cleaner + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + cp ../../es-index-cleaner-upstream/04-assert.yaml ../../es-index-cleaner-upstream/README.md . + render_install_jaeger test-es-index-cleaner-with-prefix production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options.es.index-prefix=""' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.enabled=false ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.numberOfDays=0 ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.esIndexCleaner.schedule="*/1 * * * *"' ./01-install.yaml + render_report_spans test-es-index-cleaner-with-prefix true 5 00 true 02 + '[' 6 -ne 6 ']' + jaeger=test-es-index-cleaner-with-prefix + is_secured=true + number_of_spans=5 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + export JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=5 + DAYS=5 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + sed 's~enabled: false~enabled: true~gi' ./01-install.yaml + CRONJOB_NAME=test-es-index-cleaner-with-prefix-es-index-cleaner + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./04-wait-es-index-cleaner.yaml + /tmp/jaeger-tests/bin/gomplate -f ./01-install.yaml -o ./05-install.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 00 06 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=00 + test_step=06 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=test-es-index-cleaner-with-prefix-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=test-es-index-cleaner-with-prefix-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./06-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./06-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.0"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.0"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.0"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.0 ++ version_ge 5.8.0 5.4 +++ echo 5.8.0 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.0 == 5.8.0 + '[' -n '' ']' + skip_test es-index-cleaner-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-index-cleaner-autoprov + '[' es-index-cleaner-autoprov '!=' _build ']' + cd .. + rm -rf es-index-cleaner-managed + warning 'es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + start_test es-multiinstance + '[' 1 -ne 1 ']' + test_name=es-multiinstance + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-multiinstance' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-multiinstance\e[0m' Rendering files for test es-multiinstance + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-multiinstance + cd es-multiinstance + jaeger_name=instance-1 + render_install_jaeger instance-1 production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=instance-1 + JAEGER_NAME=instance-1 + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f ./03-create-second-instance.yaml.template -o 03-create-second-instance.yaml + '[' true = true ']' + skip_test es-rollover-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-rollover-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-multiinstance + '[' es-multiinstance '!=' _build ']' + cd .. + rm -rf es-rollover-upstream + warning 'es-rollover-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_rollover -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-rollover-autoprov + '[' 1 -ne 1 ']' + test_name=es-rollover-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-rollover-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-rollover-autoprov\e[0m' Rendering files for test es-rollover-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-rollover-autoprov + cd es-rollover-autoprov + cp ../../es-rollover-upstream/05-assert.yaml ../../es-rollover-upstream/05-install.yaml ../../es-rollover-upstream/README.md . + jaeger_name=my-jaeger + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_report_spans my-jaeger true 2 00 true 02 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 00 03 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=00 + test_step=03 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./03-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./03-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 01 04 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=01 + test_step=04 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=01 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./04-check-indices.yaml + JOB_NUMBER=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./04-assert.yaml + render_report_spans my-jaeger true 2 02 true 06 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=02 + ensure_reported_spans=true + test_step=06 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=02 + JOB_NUMBER=02 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./06-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./06-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 02 07 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=02 + test_step=07 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=02 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./07-check-indices.yaml + JOB_NUMBER=02 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./07-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' 03 08 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + job_number=03 + test_step=08 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=03 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./08-check-indices.yaml + JOB_NUMBER=03 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./08-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' 04 09 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + job_number=04 + test_step=09 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=04 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./09-check-indices.yaml + JOB_NUMBER=04 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./09-assert.yaml + render_report_spans my-jaeger true 2 03 true 10 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=03 + ensure_reported_spans=true + test_step=10 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=03 + JOB_NUMBER=03 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./10-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./10-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + CRONJOB_NAME=my-jaeger-es-rollover + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./11-wait-rollover.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-000002'\'',' 05 11 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-000002'\'',' + job_number=05 + test_step=11 + escape_command ''\''--name'\'', '\''jaeger-span-000002'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-000002'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-000002'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-000002'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=05 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./11-check-indices.yaml + JOB_NUMBER=05 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./11-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' 06 12 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + job_number=06 + test_step=12 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=06 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./12-check-indices.yaml + JOB_NUMBER=06 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./12-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.0"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.0"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.0"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.0 ++ version_ge 5.8.0 5.4 +++ echo 5.8.0 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.0 == 5.8.0 + '[' -n '' ']' + skip_test es-rollover-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-rollover-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-rollover-autoprov + '[' es-rollover-autoprov '!=' _build ']' + cd .. + rm -rf es-rollover-managed + warning 'es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + skip_test es-spark-dependencies 'This test is not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=es-spark-dependencies + message='This test is not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + rm -rf es-spark-dependencies + warning 'es-spark-dependencies: This test is not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-spark-dependencies: This test is not supported in OpenShift\e[0m' WAR: es-spark-dependencies: This test is not supported in OpenShift + [[ true = true ]] + [[ false = false ]] + start_test es-streaming-autoprovisioned + '[' 1 -ne 1 ']' + test_name=es-streaming-autoprovisioned + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-streaming-autoprovisioned' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-streaming-autoprovisioned\e[0m' Rendering files for test es-streaming-autoprovisioned + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-streaming-autoprovisioned + cd es-streaming-autoprovisioned + jaeger_name=auto-provisioned + render_assert_kafka true auto-provisioned 00 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provisioned + test_step=00 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./1-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_smoke_test auto-provisioned true 04 + '[' 3 -ne 3 ']' + jaeger=auto-provisioned + is_secured=true + test_step=04 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://auto-provisioned-query:443 + JAEGER_QUERY_ENDPOINT=https://auto-provisioned-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://auto-provisioned-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://auto-provisioned-collector-headless:14268 + export JAEGER_NAME=auto-provisioned + JAEGER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./04-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running elasticsearch E2E tests' Running elasticsearch E2E tests + cd tests/e2e/elasticsearch/_build + set +e + KUBECONFIG=/tmp/kubeconfig-4018005699 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 8 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/es-from-aio-to-production === PAUSE kuttl/harness/es-from-aio-to-production === RUN kuttl/harness/es-increasing-replicas === PAUSE kuttl/harness/es-increasing-replicas === RUN kuttl/harness/es-index-cleaner-autoprov === PAUSE kuttl/harness/es-index-cleaner-autoprov === RUN kuttl/harness/es-multiinstance === PAUSE kuttl/harness/es-multiinstance === RUN kuttl/harness/es-rollover-autoprov === PAUSE kuttl/harness/es-rollover-autoprov === RUN kuttl/harness/es-simple-prod === PAUSE kuttl/harness/es-simple-prod === RUN kuttl/harness/es-streaming-autoprovisioned === PAUSE kuttl/harness/es-streaming-autoprovisioned === CONT kuttl/harness/artifacts logger.go:42: 14:03:13 | artifacts | Creating namespace: kuttl-test-immortal-koala logger.go:42: 14:03:13 | artifacts | artifacts events from ns kuttl-test-immortal-koala: logger.go:42: 14:03:13 | artifacts | Deleting namespace: kuttl-test-immortal-koala === CONT kuttl/harness/es-multiinstance logger.go:42: 14:03:19 | es-multiinstance | Ignoring 03-create-second-instance.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:03:19 | es-multiinstance | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:03:19 | es-multiinstance | Creating namespace: kuttl-test-noted-elephant logger.go:42: 14:03:19 | es-multiinstance/0-clear-namespace | starting test step 0-clear-namespace logger.go:42: 14:03:19 | es-multiinstance/0-clear-namespace | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --ignore-not-found=true] logger.go:42: 14:03:19 | es-multiinstance/0-clear-namespace | test step completed 0-clear-namespace logger.go:42: 14:03:19 | es-multiinstance/1-install | starting test step 1-install logger.go:42: 14:03:19 | es-multiinstance/1-install | Jaeger:kuttl-test-noted-elephant/instance-1 created logger.go:42: 14:04:37 | es-multiinstance/1-install | test step completed 1-install logger.go:42: 14:04:37 | es-multiinstance/2-create-namespace | starting test step 2-create-namespace logger.go:42: 14:04:37 | es-multiinstance/2-create-namespace | running command: [sh -c kubectl create namespace jaeger-e2e-multiinstance-test] logger.go:42: 14:04:37 | es-multiinstance/2-create-namespace | namespace/jaeger-e2e-multiinstance-test created logger.go:42: 14:04:37 | es-multiinstance/2-create-namespace | test step completed 2-create-namespace logger.go:42: 14:04:37 | es-multiinstance/3-create-second-instance | starting test step 3-create-second-instance logger.go:42: 14:04:37 | es-multiinstance/3-create-second-instance | running command: [sh -c kubectl apply -f ./01-install.yaml -n jaeger-e2e-multiinstance-test] logger.go:42: 14:04:40 | es-multiinstance/3-create-second-instance | jaeger.jaegertracing.io/instance-1 created logger.go:42: 14:04:40 | es-multiinstance/3-create-second-instance | running command: [sh -c /tmp/jaeger-tests/bin/kubectl-kuttl assert ./01-assert.yaml -n jaeger-e2e-multiinstance-test --timeout 1000] logger.go:42: 14:05:31 | es-multiinstance/3-create-second-instance | assert is valid logger.go:42: 14:05:31 | es-multiinstance/3-create-second-instance | test step completed 3-create-second-instance logger.go:42: 14:05:31 | es-multiinstance/4-check-secrets | starting test step 4-check-secrets logger.go:42: 14:05:31 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n $NAMESPACE > secret1] logger.go:42: 14:05:31 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n jaeger-e2e-multiinstance-test > secret2] logger.go:42: 14:05:31 | es-multiinstance/4-check-secrets | running command: [sh -c cmp --silent secret1 secret2 || exit 0] logger.go:42: 14:05:31 | es-multiinstance/4-check-secrets | test step completed 4-check-secrets logger.go:42: 14:05:31 | es-multiinstance/5-delete | starting test step 5-delete logger.go:42: 14:05:31 | es-multiinstance/5-delete | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --wait=false] logger.go:42: 14:05:31 | es-multiinstance/5-delete | namespace "jaeger-e2e-multiinstance-test" deleted logger.go:42: 14:05:31 | es-multiinstance/5-delete | test step completed 5-delete logger.go:42: 14:05:31 | es-multiinstance | es-multiinstance events from ns kuttl-test-noted-elephant: logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:26 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestnotedelephantinstance1-1-ddb48cc8 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestnotedelephantinstance1-1-ddb48cjv79w replicaset-controller logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnotedelephantinstance1-1-ddb48cjv79w Binding Scheduled Successfully assigned kuttl-test-noted-elephant/elasticsearch-cdm-kuttltestnotedelephantinstance1-1-ddb48cjv79w to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnotedelephantinstance1-1-ddb48cjv79w AddedInterface Add eth0 [10.128.2.16/23] from ovn-kubernetes logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnotedelephantinstance1-1-ddb48cjv79w.spec.containers{elasticsearch} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:56ea62bfb0ca36e19a7b21aff3676e49511f05f72da5e76d6427fd8240f328a8" kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:26 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestnotedelephantinstance1-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestnotedelephantinstance1-1-ddb48cc8 to 1 deployment-controller logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:26 +0000 UTC Normal Pod instance-1-collector-899cdd544-2cx4m Binding Scheduled Successfully assigned kuttl-test-noted-elephant/instance-1-collector-899cdd544-2cx4m to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:26 +0000 UTC Normal Pod instance-1-collector-899cdd544-2cx4m AddedInterface Add eth0 [10.129.2.23/23] from ovn-kubernetes logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:26 +0000 UTC Normal Pod instance-1-collector-899cdd544-2cx4m.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:26 +0000 UTC Normal ReplicaSet.apps instance-1-collector-899cdd544 SuccessfulCreate Created pod: instance-1-collector-899cdd544-2cx4m replicaset-controller logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:26 +0000 UTC Normal Deployment.apps instance-1-collector ScalingReplicaSet Scaled up replica set instance-1-collector-899cdd544 to 1 deployment-controller logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:26 +0000 UTC Normal Pod instance-1-query-89995f6cb-9q2r4 Binding Scheduled Successfully assigned kuttl-test-noted-elephant/instance-1-query-89995f6cb-9q2r4 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:26 +0000 UTC Normal Pod instance-1-query-89995f6cb-9q2r4 AddedInterface Add eth0 [10.129.2.24/23] from ovn-kubernetes logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:26 +0000 UTC Normal Pod instance-1-query-89995f6cb-9q2r4.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:26 +0000 UTC Normal ReplicaSet.apps instance-1-query-89995f6cb SuccessfulCreate Created pod: instance-1-query-89995f6cb-9q2r4 replicaset-controller logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:26 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-89995f6cb to 1 deployment-controller logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:29 +0000 UTC Normal Pod instance-1-collector-899cdd544-2cx4m.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" in 2.784s (2.784s including waiting) kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:29 +0000 UTC Normal Pod instance-1-collector-899cdd544-2cx4m.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:29 +0000 UTC Normal Pod instance-1-collector-899cdd544-2cx4m.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:30 +0000 UTC Normal Pod instance-1-query-89995f6cb-9q2r4.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" in 3.377s (3.377s including waiting) kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:30 +0000 UTC Normal Pod instance-1-query-89995f6cb-9q2r4.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:30 +0000 UTC Normal Pod instance-1-query-89995f6cb-9q2r4.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:30 +0000 UTC Normal Pod instance-1-query-89995f6cb-9q2r4.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:30 +0000 UTC Normal Pod instance-1-query-89995f6cb-9q2r4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:30 +0000 UTC Normal Pod instance-1-query-89995f6cb-9q2r4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:30 +0000 UTC Normal Pod instance-1-query-89995f6cb-9q2r4.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:31 +0000 UTC Warning Pod instance-1-collector-899cdd544-2cx4m.spec.containers{jaeger-collector} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:35 +0000 UTC Normal Pod instance-1-collector-899cdd544-2cx4m.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" already present on machine kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:37 +0000 UTC Normal Pod instance-1-query-89995f6cb-9q2r4.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" in 6.797s (6.797s including waiting) kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:37 +0000 UTC Normal Pod instance-1-query-89995f6cb-9q2r4.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:37 +0000 UTC Normal Pod instance-1-query-89995f6cb-9q2r4.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:37 +0000 UTC Normal Pod instance-1-query-89995f6cb-9q2r4.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:38 +0000 UTC Warning Pod instance-1-query-89995f6cb-9q2r4.spec.containers{jaeger-query} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:41 +0000 UTC Warning Pod instance-1-collector-899cdd544-2cx4m.spec.containers{jaeger-collector} BackOff Back-off restarting failed container jaeger-collector in pod instance-1-collector-899cdd544-2cx4m_kuttl-test-noted-elephant(859fcd33-d969-4f27-a65d-55fd4bee7e66) kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:42 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:43 +0000 UTC Warning Pod instance-1-query-89995f6cb-9q2r4.spec.containers{jaeger-query} BackOff Back-off restarting failed container jaeger-query in pod instance-1-query-89995f6cb-9q2r4_kuttl-test-noted-elephant(e703fa8d-184e-42a6-a9f7-3ec588b20d7a) kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnotedelephantinstance1-1-ddb48cjv79w.spec.containers{elasticsearch} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:56ea62bfb0ca36e19a7b21aff3676e49511f05f72da5e76d6427fd8240f328a8" in 23.062s (23.062s including waiting) kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnotedelephantinstance1-1-ddb48cjv79w.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnotedelephantinstance1-1-ddb48cjv79w.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnotedelephantinstance1-1-ddb48cjv79w.spec.containers{proxy} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:79427bea6b5c37894b9782c3821d8b9074838e606daa4a743b2ae060856fa98a" kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnotedelephantinstance1-1-ddb48cjv79w.spec.containers{proxy} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:79427bea6b5c37894b9782c3821d8b9074838e606daa4a743b2ae060856fa98a" in 6.875s (6.875s including waiting) kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnotedelephantinstance1-1-ddb48cjv79w.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnotedelephantinstance1-1-ddb48cjv79w.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:03:59 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestnotedelephantinstance1-1-ddb48cjv79w.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:06 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestnotedelephantinstance1-1-ddb48cjv79w.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:41 +0000 UTC Normal Pod instance-1-query-789f66c8dc-xbzxz Binding Scheduled Successfully assigned kuttl-test-noted-elephant/instance-1-query-789f66c8dc-xbzxz to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:41 +0000 UTC Normal ReplicaSet.apps instance-1-query-789f66c8dc SuccessfulCreate Created pod: instance-1-query-789f66c8dc-xbzxz replicaset-controller logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:41 +0000 UTC Normal ReplicaSet.apps instance-1-query-89995f6cb SuccessfulDelete Deleted pod: instance-1-query-89995f6cb-9q2r4 replicaset-controller logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:41 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled down replica set instance-1-query-89995f6cb to 0 from 1 deployment-controller logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:41 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-789f66c8dc to 1 deployment-controller logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:42 +0000 UTC Normal Pod instance-1-query-789f66c8dc-xbzxz AddedInterface Add eth0 [10.129.2.25/23] from ovn-kubernetes logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:42 +0000 UTC Normal Pod instance-1-query-789f66c8dc-xbzxz.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:42 +0000 UTC Normal Pod instance-1-query-789f66c8dc-xbzxz.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:42 +0000 UTC Normal Pod instance-1-query-789f66c8dc-xbzxz.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:42 +0000 UTC Normal Pod instance-1-query-789f66c8dc-xbzxz.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:42 +0000 UTC Normal Pod instance-1-query-789f66c8dc-xbzxz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:42 +0000 UTC Normal Pod instance-1-query-789f66c8dc-xbzxz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:42 +0000 UTC Normal Pod instance-1-query-789f66c8dc-xbzxz.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:42 +0000 UTC Normal Pod instance-1-query-789f66c8dc-xbzxz.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:42 +0000 UTC Normal Pod instance-1-query-789f66c8dc-xbzxz.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:48 +0000 UTC Normal Pod instance-1-query-789f66c8dc-xbzxz.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:48 +0000 UTC Normal Pod instance-1-query-789f66c8dc-xbzxz.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:48 +0000 UTC Normal Pod instance-1-query-789f66c8dc-xbzxz.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:48 +0000 UTC Normal ReplicaSet.apps instance-1-query-789f66c8dc SuccessfulDelete Deleted pod: instance-1-query-789f66c8dc-xbzxz replicaset-controller logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:48 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled down replica set instance-1-query-789f66c8dc to 0 from 1 deployment-controller logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:49 +0000 UTC Normal Pod instance-1-query-89995f6cb-8zh7s Binding Scheduled Successfully assigned kuttl-test-noted-elephant/instance-1-query-89995f6cb-8zh7s to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:49 +0000 UTC Normal Pod instance-1-query-89995f6cb-8zh7s AddedInterface Add eth0 [10.129.2.26/23] from ovn-kubernetes logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:49 +0000 UTC Normal Pod instance-1-query-89995f6cb-8zh7s.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:49 +0000 UTC Normal Pod instance-1-query-89995f6cb-8zh7s.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:49 +0000 UTC Normal Pod instance-1-query-89995f6cb-8zh7s.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:49 +0000 UTC Normal Pod instance-1-query-89995f6cb-8zh7s.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:49 +0000 UTC Normal Pod instance-1-query-89995f6cb-8zh7s.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:49 +0000 UTC Normal Pod instance-1-query-89995f6cb-8zh7s.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:49 +0000 UTC Normal Pod instance-1-query-89995f6cb-8zh7s.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:49 +0000 UTC Normal Pod instance-1-query-89995f6cb-8zh7s.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:49 +0000 UTC Normal ReplicaSet.apps instance-1-query-89995f6cb SuccessfulCreate Created pod: instance-1-query-89995f6cb-8zh7s replicaset-controller logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:49 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-89995f6cb to 1 from 0 deployment-controller logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:50 +0000 UTC Normal Pod instance-1-query-89995f6cb-8zh7s.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:04:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 14:05:31 | es-multiinstance | 2023-11-27 14:05:28 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod instance-1-collector-899cdd544-2cx4m horizontal-pod-autoscaler logger.go:42: 14:05:31 | es-multiinstance | Deleting namespace: kuttl-test-noted-elephant === CONT kuttl/harness/es-streaming-autoprovisioned logger.go:42: 14:05:38 | es-streaming-autoprovisioned | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:05:38 | es-streaming-autoprovisioned | Creating namespace: kuttl-test-fine-monitor logger.go:42: 14:05:38 | es-streaming-autoprovisioned/0-install | starting test step 0-install logger.go:42: 14:05:38 | es-streaming-autoprovisioned/0-install | Jaeger:kuttl-test-fine-monitor/auto-provisioned created logger.go:42: 14:06:52 | es-streaming-autoprovisioned/0-install | test step completed 0-install logger.go:42: 14:06:52 | es-streaming-autoprovisioned/1- | starting test step 1- logger.go:42: 14:07:23 | es-streaming-autoprovisioned/1- | test step completed 1- logger.go:42: 14:07:23 | es-streaming-autoprovisioned/2- | starting test step 2- logger.go:42: 14:07:47 | es-streaming-autoprovisioned/2- | test step completed 2- logger.go:42: 14:07:47 | es-streaming-autoprovisioned/3- | starting test step 3- logger.go:42: 14:08:01 | es-streaming-autoprovisioned/3- | test step completed 3- logger.go:42: 14:08:01 | es-streaming-autoprovisioned/4-smoke-test | starting test step 4-smoke-test logger.go:42: 14:08:01 | es-streaming-autoprovisioned/4-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE auto-provisioned /dev/null] logger.go:42: 14:08:04 | es-streaming-autoprovisioned/4-smoke-test | Warning: resource jaegers/auto-provisioned is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:08:12 | es-streaming-autoprovisioned/4-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JAEGER_COLLECTOR_ENDPOINT=http://auto-provisioned-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://auto-provisioned-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 14:08:13 | es-streaming-autoprovisioned/4-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:08:13 | es-streaming-autoprovisioned/4-smoke-test | job.batch/report-span created logger.go:42: 14:08:13 | es-streaming-autoprovisioned/4-smoke-test | job.batch/check-span created logger.go:42: 14:08:26 | es-streaming-autoprovisioned/4-smoke-test | test step completed 4-smoke-test logger.go:42: 14:08:26 | es-streaming-autoprovisioned | es-streaming-autoprovisioned events from ns kuttl-test-fine-monitor: logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:05:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfinemonitorautoprovisioned-1-6bdkbvg Binding Scheduled Successfully assigned kuttl-test-fine-monitor/elasticsearch-cdm-kuttltestfinemonitorautoprovisioned-1-6bdkbvg to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:05:44 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestfinemonitorautoprovisioned-1-6bf54c58d SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestfinemonitorautoprovisioned-1-6bdkbvg replicaset-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:05:44 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestfinemonitorautoprovisioned-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestfinemonitorautoprovisioned-1-6bf54c58d to 1 deployment-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:05:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfinemonitorautoprovisioned-1-6bdkbvg AddedInterface Add eth0 [10.128.2.17/23] from ovn-kubernetes logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:05:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfinemonitorautoprovisioned-1-6bdkbvg.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:56ea62bfb0ca36e19a7b21aff3676e49511f05f72da5e76d6427fd8240f328a8" already present on machine kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:05:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfinemonitorautoprovisioned-1-6bdkbvg.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:05:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfinemonitorautoprovisioned-1-6bdkbvg.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:05:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfinemonitorautoprovisioned-1-6bdkbvg.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:79427bea6b5c37894b9782c3821d8b9074838e606daa4a743b2ae060856fa98a" already present on machine kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:05:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfinemonitorautoprovisioned-1-6bdkbvg.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:05:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfinemonitorautoprovisioned-1-6bdkbvg.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:05:55 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfinemonitorautoprovisioned-1-6bdkbvg.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:00 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfinemonitorautoprovisioned-1-6bdkbvg.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:13 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-zookeeper NoPods No matching pods found controllermanager logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:13 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:14 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:14 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-fine-monitor/data-auto-provisioned-zookeeper-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-684659d4f7-2wwfg_180f48e7-ecaa-4b6b-8bcf-6bd94c7ad5a7 logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:17 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ProvisioningSucceeded Successfully provisioned volume pvc-15948b58-7059-4ef0-93a4-7b8a7c159c50 ebs.csi.aws.com_aws-ebs-csi-driver-controller-684659d4f7-2wwfg_180f48e7-ecaa-4b6b-8bcf-6bd94c7ad5a7 logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:18 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 Binding Scheduled Successfully assigned kuttl-test-fine-monitor/auto-provisioned-zookeeper-0 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:20 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-15948b58-7059-4ef0-93a4-7b8a7c159c50" attachdetach-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:22 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 AddedInterface Add eth0 [10.129.2.29/23] from ovn-kubernetes logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:22 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:0bcdd55f01638f650ed69ebdf5f8a7291e103805b8cbb34013ced88e46e0678c" kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:31 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:0bcdd55f01638f650ed69ebdf5f8a7291e103805b8cbb34013ced88e46e0678c" in 9.534s (9.534s including waiting) kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:31 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Created Created container zookeeper kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:31 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Started Started container zookeeper kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:53 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:54 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-kafka NoPods No matching pods found controllermanager logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:54 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:54 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-fine-monitor/data-0-auto-provisioned-kafka-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-684659d4f7-2wwfg_180f48e7-ecaa-4b6b-8bcf-6bd94c7ad5a7 logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:57 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ProvisioningSucceeded Successfully provisioned volume pvc-b9890857-69ff-4a71-92dd-96095a9e2f60 ebs.csi.aws.com_aws-ebs-csi-driver-controller-684659d4f7-2wwfg_180f48e7-ecaa-4b6b-8bcf-6bd94c7ad5a7 logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:06:58 +0000 UTC Normal Pod auto-provisioned-kafka-0 Binding Scheduled Successfully assigned kuttl-test-fine-monitor/auto-provisioned-kafka-0 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:00 +0000 UTC Normal Pod auto-provisioned-kafka-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-b9890857-69ff-4a71-92dd-96095a9e2f60" attachdetach-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:02 +0000 UTC Normal Pod auto-provisioned-kafka-0 AddedInterface Add eth0 [10.129.2.30/23] from ovn-kubernetes logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:03 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulled Container image "registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:0bcdd55f01638f650ed69ebdf5f8a7291e103805b8cbb34013ced88e46e0678c" already present on machine kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:03 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Created Created container kafka kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:03 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Started Started container kafka kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:25 +0000 UTC Normal Pod auto-provisioned-entity-operator-587b68d6b4-km69h Binding Scheduled Successfully assigned kuttl-test-fine-monitor/auto-provisioned-entity-operator-587b68d6b4-km69h to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:25 +0000 UTC Normal Pod auto-provisioned-entity-operator-587b68d6b4-km69h AddedInterface Add eth0 [10.129.2.31/23] from ovn-kubernetes logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:25 +0000 UTC Normal Pod auto-provisioned-entity-operator-587b68d6b4-km69h.spec.containers{topic-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:464b04e622e0b3472e8a1e1ce8a2efd32cf27fc2056d3d589bfe6b5f9ac0bf4e" already present on machine kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:25 +0000 UTC Normal Pod auto-provisioned-entity-operator-587b68d6b4-km69h.spec.containers{topic-operator} Created Created container topic-operator kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:25 +0000 UTC Normal Pod auto-provisioned-entity-operator-587b68d6b4-km69h.spec.containers{topic-operator} Started Started container topic-operator kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:25 +0000 UTC Normal Pod auto-provisioned-entity-operator-587b68d6b4-km69h.spec.containers{user-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:464b04e622e0b3472e8a1e1ce8a2efd32cf27fc2056d3d589bfe6b5f9ac0bf4e" already present on machine kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:25 +0000 UTC Normal Pod auto-provisioned-entity-operator-587b68d6b4-km69h.spec.containers{user-operator} Created Created container user-operator kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:25 +0000 UTC Normal Pod auto-provisioned-entity-operator-587b68d6b4-km69h.spec.containers{user-operator} Started Started container user-operator kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:25 +0000 UTC Normal Pod auto-provisioned-entity-operator-587b68d6b4-km69h.spec.containers{tls-sidecar} Pulled Container image "registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:0bcdd55f01638f650ed69ebdf5f8a7291e103805b8cbb34013ced88e46e0678c" already present on machine kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:25 +0000 UTC Normal Pod auto-provisioned-entity-operator-587b68d6b4-km69h.spec.containers{tls-sidecar} Created Created container tls-sidecar kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:25 +0000 UTC Normal Pod auto-provisioned-entity-operator-587b68d6b4-km69h.spec.containers{tls-sidecar} Started Started container tls-sidecar kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:25 +0000 UTC Normal ReplicaSet.apps auto-provisioned-entity-operator-587b68d6b4 SuccessfulCreate Created pod: auto-provisioned-entity-operator-587b68d6b4-km69h replicaset-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:25 +0000 UTC Normal Deployment.apps auto-provisioned-entity-operator ScalingReplicaSet Scaled up replica set auto-provisioned-entity-operator-587b68d6b4 to 1 deployment-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:48 +0000 UTC Normal Pod auto-provisioned-collector-f8dc59f5c-xsnnr Binding Scheduled Successfully assigned kuttl-test-fine-monitor/auto-provisioned-collector-f8dc59f5c-xsnnr to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:48 +0000 UTC Normal ReplicaSet.apps auto-provisioned-collector-f8dc59f5c SuccessfulCreate Created pod: auto-provisioned-collector-f8dc59f5c-xsnnr replicaset-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:48 +0000 UTC Normal Deployment.apps auto-provisioned-collector ScalingReplicaSet Scaled up replica set auto-provisioned-collector-f8dc59f5c to 1 deployment-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:48 +0000 UTC Normal Deployment.apps auto-provisioned-ingester ScalingReplicaSet Scaled up replica set auto-provisioned-ingester-5ff8b59b99 to 1 deployment-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:48 +0000 UTC Normal Pod auto-provisioned-query-79b5bf7547-jns6j Binding Scheduled Successfully assigned kuttl-test-fine-monitor/auto-provisioned-query-79b5bf7547-jns6j to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:48 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-79b5bf7547 SuccessfulCreate Created pod: auto-provisioned-query-79b5bf7547-jns6j replicaset-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:48 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled up replica set auto-provisioned-query-79b5bf7547 to 1 deployment-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:49 +0000 UTC Normal Pod auto-provisioned-collector-f8dc59f5c-xsnnr AddedInterface Add eth0 [10.131.0.23/23] from ovn-kubernetes logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:49 +0000 UTC Normal Pod auto-provisioned-collector-f8dc59f5c-xsnnr.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:49 +0000 UTC Normal Pod auto-provisioned-ingester-5ff8b59b99-tcvzl Binding Scheduled Successfully assigned kuttl-test-fine-monitor/auto-provisioned-ingester-5ff8b59b99-tcvzl to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:49 +0000 UTC Normal Pod auto-provisioned-ingester-5ff8b59b99-tcvzl AddedInterface Add eth0 [10.131.0.25/23] from ovn-kubernetes logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:49 +0000 UTC Normal Pod auto-provisioned-ingester-5ff8b59b99-tcvzl.spec.containers{jaeger-ingester} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:109a3676bfaccda0021c57bbe82bceed5140faaedb8bad4d7d2be7cd660de039" kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:49 +0000 UTC Normal ReplicaSet.apps auto-provisioned-ingester-5ff8b59b99 SuccessfulCreate Created pod: auto-provisioned-ingester-5ff8b59b99-tcvzl replicaset-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:49 +0000 UTC Normal Pod auto-provisioned-query-79b5bf7547-jns6j AddedInterface Add eth0 [10.131.0.24/23] from ovn-kubernetes logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:49 +0000 UTC Normal Pod auto-provisioned-query-79b5bf7547-jns6j.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:56 +0000 UTC Normal Pod auto-provisioned-collector-f8dc59f5c-xsnnr.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" in 7.43s (7.43s including waiting) kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:56 +0000 UTC Normal Pod auto-provisioned-collector-f8dc59f5c-xsnnr.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:56 +0000 UTC Normal Pod auto-provisioned-collector-f8dc59f5c-xsnnr.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:56 +0000 UTC Normal Pod auto-provisioned-ingester-5ff8b59b99-tcvzl.spec.containers{jaeger-ingester} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:109a3676bfaccda0021c57bbe82bceed5140faaedb8bad4d7d2be7cd660de039" in 7.319s (7.319s including waiting) kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:56 +0000 UTC Normal Pod auto-provisioned-ingester-5ff8b59b99-tcvzl.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:56 +0000 UTC Normal Pod auto-provisioned-ingester-5ff8b59b99-tcvzl.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:56 +0000 UTC Normal Pod auto-provisioned-query-79b5bf7547-jns6j.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" in 7.367s (7.367s including waiting) kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:56 +0000 UTC Normal Pod auto-provisioned-query-79b5bf7547-jns6j.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:56 +0000 UTC Normal Pod auto-provisioned-query-79b5bf7547-jns6j.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:56 +0000 UTC Normal Pod auto-provisioned-query-79b5bf7547-jns6j.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:57 +0000 UTC Normal Pod auto-provisioned-query-79b5bf7547-jns6j.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:57 +0000 UTC Normal Pod auto-provisioned-query-79b5bf7547-jns6j.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:57 +0000 UTC Normal Pod auto-provisioned-query-79b5bf7547-jns6j.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:59 +0000 UTC Normal Pod auto-provisioned-query-79b5bf7547-jns6j.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" in 2.459s (2.459s including waiting) kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:59 +0000 UTC Normal Pod auto-provisioned-query-79b5bf7547-jns6j.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:07:59 +0000 UTC Normal Pod auto-provisioned-query-79b5bf7547-jns6j.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:04 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:04 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:04 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:04 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-ingester FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:04 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-ingester FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:04 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-ingester FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:05 +0000 UTC Normal Pod auto-provisioned-query-79b5bf7547-jns6j.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:05 +0000 UTC Normal Pod auto-provisioned-query-79b5bf7547-jns6j.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:05 +0000 UTC Normal Pod auto-provisioned-query-79b5bf7547-jns6j.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:05 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-79b5bf7547 SuccessfulDelete Deleted pod: auto-provisioned-query-79b5bf7547-jns6j replicaset-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:05 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled down replica set auto-provisioned-query-79b5bf7547 to 0 from 1 deployment-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:06 +0000 UTC Normal Pod auto-provisioned-query-6f6fd97d7d-hxfx4 Binding Scheduled Successfully assigned kuttl-test-fine-monitor/auto-provisioned-query-6f6fd97d7d-hxfx4 to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:06 +0000 UTC Normal Pod auto-provisioned-query-6f6fd97d7d-hxfx4 AddedInterface Add eth0 [10.131.0.26/23] from ovn-kubernetes logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:06 +0000 UTC Normal Pod auto-provisioned-query-6f6fd97d7d-hxfx4.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:06 +0000 UTC Normal Pod auto-provisioned-query-6f6fd97d7d-hxfx4.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:06 +0000 UTC Normal Pod auto-provisioned-query-6f6fd97d7d-hxfx4.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:06 +0000 UTC Normal Pod auto-provisioned-query-6f6fd97d7d-hxfx4.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:06 +0000 UTC Normal Pod auto-provisioned-query-6f6fd97d7d-hxfx4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:06 +0000 UTC Normal Pod auto-provisioned-query-6f6fd97d7d-hxfx4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:06 +0000 UTC Normal Pod auto-provisioned-query-6f6fd97d7d-hxfx4.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:06 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-6f6fd97d7d SuccessfulCreate Created pod: auto-provisioned-query-6f6fd97d7d-hxfx4 replicaset-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:06 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled up replica set auto-provisioned-query-6f6fd97d7d to 1 deployment-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:07 +0000 UTC Normal Pod auto-provisioned-query-6f6fd97d7d-hxfx4.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:07 +0000 UTC Normal Pod auto-provisioned-query-6f6fd97d7d-hxfx4.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:13 +0000 UTC Normal Pod check-span-vc6ts Binding Scheduled Successfully assigned kuttl-test-fine-monitor/check-span-vc6ts to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:13 +0000 UTC Normal Pod check-span-vc6ts AddedInterface Add eth0 [10.131.0.27/23] from ovn-kubernetes logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:13 +0000 UTC Normal Pod check-span-vc6ts.spec.containers{asserts-container} Pulling Pulling image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:13 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-vc6ts job-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:13 +0000 UTC Normal Pod report-span-ps7xp Binding Scheduled Successfully assigned kuttl-test-fine-monitor/report-span-ps7xp to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:13 +0000 UTC Normal Pod report-span-ps7xp AddedInterface Add eth0 [10.129.2.32/23] from ovn-kubernetes logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:13 +0000 UTC Normal Pod report-span-ps7xp.spec.containers{report-span} Pulling Pulling image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:13 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-ps7xp job-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:15 +0000 UTC Normal Pod check-span-vc6ts.spec.containers{asserts-container} Pulled Successfully pulled image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" in 1.745s (1.745s including waiting) kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:15 +0000 UTC Normal Pod check-span-vc6ts.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:15 +0000 UTC Normal Pod check-span-vc6ts.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:15 +0000 UTC Normal Pod report-span-ps7xp.spec.containers{report-span} Pulled Successfully pulled image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" in 1.789s (1.789s including waiting) kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:15 +0000 UTC Normal Pod report-span-ps7xp.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:15 +0000 UTC Normal Pod report-span-ps7xp.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:08:26 | es-streaming-autoprovisioned | 2023-11-27 14:08:26 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:08:26 | es-streaming-autoprovisioned | Deleting namespace: kuttl-test-fine-monitor === CONT kuttl/harness/es-simple-prod logger.go:42: 14:09:07 | es-simple-prod | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:09:07 | es-simple-prod | Creating namespace: kuttl-test-simple-orca logger.go:42: 14:09:07 | es-simple-prod | es-simple-prod events from ns kuttl-test-simple-orca: logger.go:42: 14:09:07 | es-simple-prod | Deleting namespace: kuttl-test-simple-orca === CONT kuttl/harness/es-rollover-autoprov logger.go:42: 14:09:13 | es-rollover-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:09:13 | es-rollover-autoprov | Creating namespace: kuttl-test-capital-shepherd logger.go:42: 14:09:13 | es-rollover-autoprov/1-install | starting test step 1-install logger.go:42: 14:09:13 | es-rollover-autoprov/1-install | Jaeger:kuttl-test-capital-shepherd/my-jaeger created logger.go:42: 14:09:48 | es-rollover-autoprov/1-install | test step completed 1-install logger.go:42: 14:09:48 | es-rollover-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 14:09:48 | es-rollover-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 14:09:50 | es-rollover-autoprov/2-report-spans | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:09:56 | es-rollover-autoprov/2-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 14:09:57 | es-rollover-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 14:09:57 | es-rollover-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 14:10:20 | es-rollover-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 14:10:20 | es-rollover-autoprov/3-check-indices | starting test step 3-check-indices logger.go:42: 14:10:20 | es-rollover-autoprov/3-check-indices | Job:kuttl-test-capital-shepherd/00-check-indices created logger.go:42: 14:10:23 | es-rollover-autoprov/3-check-indices | test step completed 3-check-indices logger.go:42: 14:10:23 | es-rollover-autoprov/4-check-indices | starting test step 4-check-indices logger.go:42: 14:10:24 | es-rollover-autoprov/4-check-indices | Job:kuttl-test-capital-shepherd/01-check-indices created logger.go:42: 14:10:28 | es-rollover-autoprov/4-check-indices | test step completed 4-check-indices logger.go:42: 14:10:28 | es-rollover-autoprov/5-install | starting test step 5-install logger.go:42: 14:10:28 | es-rollover-autoprov/5-install | Jaeger:kuttl-test-capital-shepherd/my-jaeger updated logger.go:42: 14:10:35 | es-rollover-autoprov/5-install | test step completed 5-install logger.go:42: 14:10:35 | es-rollover-autoprov/6-report-spans | starting test step 6-report-spans logger.go:42: 14:10:35 | es-rollover-autoprov/6-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 14:10:42 | es-rollover-autoprov/6-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JOB_NUMBER=02 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-02-job.yaml] logger.go:42: 14:10:43 | es-rollover-autoprov/6-report-spans | running command: [sh -c kubectl apply -f report-span-02-job.yaml -n $NAMESPACE] logger.go:42: 14:10:43 | es-rollover-autoprov/6-report-spans | job.batch/02-report-span created logger.go:42: 14:11:06 | es-rollover-autoprov/6-report-spans | test step completed 6-report-spans logger.go:42: 14:11:06 | es-rollover-autoprov/7-check-indices | starting test step 7-check-indices logger.go:42: 14:11:07 | es-rollover-autoprov/7-check-indices | Job:kuttl-test-capital-shepherd/02-check-indices created logger.go:42: 14:11:10 | es-rollover-autoprov/7-check-indices | test step completed 7-check-indices logger.go:42: 14:11:10 | es-rollover-autoprov/8-check-indices | starting test step 8-check-indices logger.go:42: 14:11:10 | es-rollover-autoprov/8-check-indices | Job:kuttl-test-capital-shepherd/03-check-indices created logger.go:42: 14:11:13 | es-rollover-autoprov/8-check-indices | test step completed 8-check-indices logger.go:42: 14:11:13 | es-rollover-autoprov/9-check-indices | starting test step 9-check-indices logger.go:42: 14:11:13 | es-rollover-autoprov/9-check-indices | Job:kuttl-test-capital-shepherd/04-check-indices created logger.go:42: 14:11:16 | es-rollover-autoprov/9-check-indices | test step completed 9-check-indices logger.go:42: 14:11:16 | es-rollover-autoprov/10-report-spans | starting test step 10-report-spans logger.go:42: 14:11:16 | es-rollover-autoprov/10-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 14:11:23 | es-rollover-autoprov/10-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JOB_NUMBER=03 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-03-job.yaml] logger.go:42: 14:11:24 | es-rollover-autoprov/10-report-spans | running command: [sh -c kubectl apply -f report-span-03-job.yaml -n $NAMESPACE] logger.go:42: 14:11:24 | es-rollover-autoprov/10-report-spans | job.batch/03-report-span created logger.go:42: 14:11:48 | es-rollover-autoprov/10-report-spans | test step completed 10-report-spans logger.go:42: 14:11:48 | es-rollover-autoprov/11-check-indices | starting test step 11-check-indices logger.go:42: 14:11:48 | es-rollover-autoprov/11-check-indices | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob my-jaeger-es-rollover --namespace $NAMESPACE] logger.go:42: 14:12:00 | es-rollover-autoprov/11-check-indices | time="2023-11-27T14:12:00Z" level=debug msg="Checking if the my-jaeger-es-rollover CronJob exists" logger.go:42: 14:12:00 | es-rollover-autoprov/11-check-indices | time="2023-11-27T14:12:00Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 14:12:00 | es-rollover-autoprov/11-check-indices | time="2023-11-27T14:12:00Z" level=info msg="Cronjob my-jaeger-es-rollover found successfully" logger.go:42: 14:12:00 | es-rollover-autoprov/11-check-indices | time="2023-11-27T14:12:00Z" level=debug msg="Waiting for the next scheduled job from my-jaeger-es-rollover cronjob" logger.go:42: 14:12:00 | es-rollover-autoprov/11-check-indices | time="2023-11-27T14:12:00Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 14:12:00 | es-rollover-autoprov/11-check-indices | time="2023-11-27T14:12:00Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 14:12:10 | es-rollover-autoprov/11-check-indices | time="2023-11-27T14:12:10Z" level=info msg="Job of owner my-jaeger-es-rollover succeeded after my-jaeger-es-rollover 10.0241429s" logger.go:42: 14:12:10 | es-rollover-autoprov/11-check-indices | Job:kuttl-test-capital-shepherd/05-check-indices created logger.go:42: 14:12:14 | es-rollover-autoprov/11-check-indices | test step completed 11-check-indices logger.go:42: 14:12:14 | es-rollover-autoprov/12-check-indices | starting test step 12-check-indices logger.go:42: 14:12:14 | es-rollover-autoprov/12-check-indices | Job:kuttl-test-capital-shepherd/06-check-indices created logger.go:42: 14:12:18 | es-rollover-autoprov/12-check-indices | test step completed 12-check-indices logger.go:42: 14:12:18 | es-rollover-autoprov | es-rollover-autoprov events from ns kuttl-test-capital-shepherd: logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:18 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestcapitalshepherdmyjaeger-1-6bcdb79484 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestcapitalshepherdmyjaeger-1-6bcdbplt26 replicaset-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:18 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcapitalshepherdmyjaeger-1-6bcdbplt26 Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/elasticsearch-cdm-kuttltestcapitalshepherdmyjaeger-1-6bcdbplt26 to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:18 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestcapitalshepherdmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestcapitalshepherdmyjaeger-1-6bcdb79484 to 1 deployment-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:19 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcapitalshepherdmyjaeger-1-6bcdbplt26 AddedInterface Add eth0 [10.128.2.18/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:19 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcapitalshepherdmyjaeger-1-6bcdbplt26.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:56ea62bfb0ca36e19a7b21aff3676e49511f05f72da5e76d6427fd8240f328a8" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:19 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcapitalshepherdmyjaeger-1-6bcdbplt26.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:19 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcapitalshepherdmyjaeger-1-6bcdbplt26.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:19 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcapitalshepherdmyjaeger-1-6bcdbplt26.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:79427bea6b5c37894b9782c3821d8b9074838e606daa4a743b2ae060856fa98a" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:19 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcapitalshepherdmyjaeger-1-6bcdbplt26.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:19 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcapitalshepherdmyjaeger-1-6bcdbplt26.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:29 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcapitalshepherdmyjaeger-1-6bcdbplt26.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:34 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcapitalshepherdmyjaeger-1-6bcdbplt26.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:45 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-twwmk Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/my-jaeger-collector-5489f5bd9b-twwmk to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:45 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5489f5bd9b SuccessfulCreate Created pod: my-jaeger-collector-5489f5bd9b-twwmk replicaset-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:45 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-5489f5bd9b to 1 deployment-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:45 +0000 UTC Normal Pod my-jaeger-query-5fcc4f8c76-6q4bb Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/my-jaeger-query-5fcc4f8c76-6q4bb to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:45 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5fcc4f8c76 SuccessfulCreate Created pod: my-jaeger-query-5fcc4f8c76-6q4bb replicaset-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:45 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-5fcc4f8c76 to 1 deployment-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:46 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-twwmk AddedInterface Add eth0 [10.129.2.33/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:46 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-twwmk.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:46 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-twwmk.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:46 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-twwmk.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:46 +0000 UTC Normal Pod my-jaeger-query-5fcc4f8c76-6q4bb AddedInterface Add eth0 [10.131.0.28/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:46 +0000 UTC Normal Pod my-jaeger-query-5fcc4f8c76-6q4bb.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:46 +0000 UTC Normal Pod my-jaeger-query-5fcc4f8c76-6q4bb.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:46 +0000 UTC Normal Pod my-jaeger-query-5fcc4f8c76-6q4bb.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:46 +0000 UTC Normal Pod my-jaeger-query-5fcc4f8c76-6q4bb.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:46 +0000 UTC Normal Pod my-jaeger-query-5fcc4f8c76-6q4bb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:46 +0000 UTC Normal Pod my-jaeger-query-5fcc4f8c76-6q4bb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:46 +0000 UTC Normal Pod my-jaeger-query-5fcc4f8c76-6q4bb.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:46 +0000 UTC Normal Pod my-jaeger-query-5fcc4f8c76-6q4bb.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:46 +0000 UTC Normal Pod my-jaeger-query-5fcc4f8c76-6q4bb.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:51 +0000 UTC Normal Pod my-jaeger-query-5fcc4f8c76-6q4bb.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:51 +0000 UTC Normal Pod my-jaeger-query-5fcc4f8c76-6q4bb.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:51 +0000 UTC Normal Pod my-jaeger-query-5fcc4f8c76-6q4bb.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:51 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5fcc4f8c76 SuccessfulDelete Deleted pod: my-jaeger-query-5fcc4f8c76-6q4bb replicaset-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:51 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-5fcc4f8c76 to 0 from 1 deployment-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:52 +0000 UTC Normal Pod my-jaeger-query-c5d766764-nx8sh Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/my-jaeger-query-c5d766764-nx8sh to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:52 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-c5d766764 SuccessfulCreate Created pod: my-jaeger-query-c5d766764-nx8sh replicaset-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:52 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-c5d766764 to 1 deployment-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:53 +0000 UTC Normal Pod my-jaeger-query-c5d766764-nx8sh AddedInterface Add eth0 [10.129.2.34/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:53 +0000 UTC Normal Pod my-jaeger-query-c5d766764-nx8sh.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:53 +0000 UTC Normal Pod my-jaeger-query-c5d766764-nx8sh.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:53 +0000 UTC Normal Pod my-jaeger-query-c5d766764-nx8sh.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:53 +0000 UTC Normal Pod my-jaeger-query-c5d766764-nx8sh.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:53 +0000 UTC Normal Pod my-jaeger-query-c5d766764-nx8sh.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:53 +0000 UTC Normal Pod my-jaeger-query-c5d766764-nx8sh.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:53 +0000 UTC Normal Pod my-jaeger-query-c5d766764-nx8sh.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:53 +0000 UTC Normal Pod my-jaeger-query-c5d766764-nx8sh.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:53 +0000 UTC Normal Pod my-jaeger-query-c5d766764-nx8sh.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:57 +0000 UTC Normal Pod 00-report-span-h297f Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/00-report-span-h297f to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:57 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-h297f job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:58 +0000 UTC Normal Pod 00-report-span-h297f AddedInterface Add eth0 [10.131.0.29/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:58 +0000 UTC Normal Pod 00-report-span-h297f.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:58 +0000 UTC Normal Pod 00-report-span-h297f.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:09:58 +0000 UTC Normal Pod 00-report-span-h297f.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:00 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:00 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:00 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:20 +0000 UTC Normal Pod 00-check-indices-jxppv Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/00-check-indices-jxppv to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:20 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-jxppv job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:20 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:21 +0000 UTC Normal Pod 00-check-indices-jxppv AddedInterface Add eth0 [10.131.0.30/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:21 +0000 UTC Normal Pod 00-check-indices-jxppv.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:21 +0000 UTC Normal Pod 00-check-indices-jxppv.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:21 +0000 UTC Normal Pod 00-check-indices-jxppv.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:23 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:24 +0000 UTC Normal Pod 01-check-indices-pshwk Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/01-check-indices-pshwk to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:24 +0000 UTC Normal Pod 01-check-indices-pshwk AddedInterface Add eth0 [10.131.0.31/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:24 +0000 UTC Normal Pod 01-check-indices-pshwk.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:24 +0000 UTC Normal Pod 01-check-indices-pshwk.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:24 +0000 UTC Normal Pod 01-check-indices-pshwk.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:24 +0000 UTC Normal Job.batch 01-check-indices SuccessfulCreate Created pod: 01-check-indices-pshwk job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:27 +0000 UTC Normal Job.batch 01-check-indices Completed Job completed job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:29 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-5pkbc Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/my-jaeger-es-rollover-create-mapping-5pkbc to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:29 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-5pkbc AddedInterface Add eth0 [10.131.0.32/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:29 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-5pkbc.spec.containers{my-jaeger-es-rollover-create-mapping} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:edda53eae462f604654d8adc71622101d2f62c9de6ecce4484523341b972723a" kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:29 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping SuccessfulCreate Created pod: my-jaeger-es-rollover-create-mapping-5pkbc job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:30 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:30 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-5489f5bd9b-twwmk horizontal-pod-autoscaler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:30 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:31 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-5pkbc.spec.containers{my-jaeger-es-rollover-create-mapping} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:edda53eae462f604654d8adc71622101d2f62c9de6ecce4484523341b972723a" in 2.316s (2.316s including waiting) kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:31 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-5pkbc.spec.containers{my-jaeger-es-rollover-create-mapping} Created Created container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:32 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-5pkbc.spec.containers{my-jaeger-es-rollover-create-mapping} Started Started container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:34 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping Completed Job completed job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:35 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-twwmk.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:35 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5489f5bd9b SuccessfulDelete Deleted pod: my-jaeger-collector-5489f5bd9b-twwmk replicaset-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:35 +0000 UTC Normal Pod my-jaeger-collector-7794fb6d5c-c5nql Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/my-jaeger-collector-7794fb6d5c-c5nql to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:35 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7794fb6d5c SuccessfulCreate Created pod: my-jaeger-collector-7794fb6d5c-c5nql replicaset-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:35 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-5489f5bd9b to 0 from 1 deployment-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:35 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7794fb6d5c to 1 deployment-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:35 +0000 UTC Normal Pod my-jaeger-query-74b4c9cf57-f2xrn Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/my-jaeger-query-74b4c9cf57-f2xrn to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:35 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-74b4c9cf57 SuccessfulCreate Created pod: my-jaeger-query-74b4c9cf57-f2xrn replicaset-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:35 +0000 UTC Normal Pod my-jaeger-query-c5d766764-nx8sh.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:35 +0000 UTC Normal Pod my-jaeger-query-c5d766764-nx8sh.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:35 +0000 UTC Normal Pod my-jaeger-query-c5d766764-nx8sh.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:35 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-c5d766764 SuccessfulDelete Deleted pod: my-jaeger-query-c5d766764-nx8sh replicaset-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:35 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-c5d766764 to 0 from 1 deployment-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:35 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-74b4c9cf57 to 1 deployment-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:36 +0000 UTC Normal Pod my-jaeger-collector-7794fb6d5c-c5nql AddedInterface Add eth0 [10.129.2.35/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:36 +0000 UTC Normal Pod my-jaeger-collector-7794fb6d5c-c5nql.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:36 +0000 UTC Normal Pod my-jaeger-collector-7794fb6d5c-c5nql.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:36 +0000 UTC Normal Pod my-jaeger-collector-7794fb6d5c-c5nql.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:36 +0000 UTC Normal Pod my-jaeger-query-74b4c9cf57-f2xrn AddedInterface Add eth0 [10.131.0.33/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:36 +0000 UTC Normal Pod my-jaeger-query-74b4c9cf57-f2xrn.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:36 +0000 UTC Normal Pod my-jaeger-query-74b4c9cf57-f2xrn.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:36 +0000 UTC Normal Pod my-jaeger-query-74b4c9cf57-f2xrn.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:36 +0000 UTC Normal Pod my-jaeger-query-74b4c9cf57-f2xrn.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:36 +0000 UTC Normal Pod my-jaeger-query-74b4c9cf57-f2xrn.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:36 +0000 UTC Normal Pod my-jaeger-query-74b4c9cf57-f2xrn.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:36 +0000 UTC Normal Pod my-jaeger-query-74b4c9cf57-f2xrn.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:36 +0000 UTC Normal Pod my-jaeger-query-74b4c9cf57-f2xrn.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:36 +0000 UTC Normal Pod my-jaeger-query-74b4c9cf57-f2xrn.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:43 +0000 UTC Normal Pod 02-report-span-zrjxg Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/02-report-span-zrjxg to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:43 +0000 UTC Normal Job.batch 02-report-span SuccessfulCreate Created pod: 02-report-span-zrjxg job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:44 +0000 UTC Normal Pod 02-report-span-zrjxg AddedInterface Add eth0 [10.129.2.36/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:44 +0000 UTC Normal Pod 02-report-span-zrjxg.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:44 +0000 UTC Normal Pod 02-report-span-zrjxg.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:10:44 +0000 UTC Normal Pod 02-report-span-zrjxg.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28351571-pbczq Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/my-jaeger-es-lookback-28351571-pbczq to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28351571-pbczq AddedInterface Add eth0 [10.129.2.38/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28351571-pbczq.spec.containers{my-jaeger-es-lookback} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:edda53eae462f604654d8adc71622101d2f62c9de6ecce4484523341b972723a" kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28351571 SuccessfulCreate Created pod: my-jaeger-es-lookback-28351571-pbczq job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28351571 cronjob-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28351571-rcsdp Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/my-jaeger-es-rollover-28351571-rcsdp to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28351571-rcsdp AddedInterface Add eth0 [10.129.2.37/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28351571-rcsdp.spec.containers{my-jaeger-es-rollover} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:edda53eae462f604654d8adc71622101d2f62c9de6ecce4484523341b972723a" kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28351571 SuccessfulCreate Created pod: my-jaeger-es-rollover-28351571-rcsdp job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28351571 cronjob-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:02 +0000 UTC Normal Pod my-jaeger-es-lookback-28351571-pbczq.spec.containers{my-jaeger-es-lookback} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:edda53eae462f604654d8adc71622101d2f62c9de6ecce4484523341b972723a" in 2.004s (2.004s including waiting) kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:02 +0000 UTC Normal Pod my-jaeger-es-lookback-28351571-pbczq.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:02 +0000 UTC Normal Pod my-jaeger-es-lookback-28351571-pbczq.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:02 +0000 UTC Normal Pod my-jaeger-es-rollover-28351571-rcsdp.spec.containers{my-jaeger-es-rollover} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:edda53eae462f604654d8adc71622101d2f62c9de6ecce4484523341b972723a" in 1.82s (1.82s including waiting) kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:02 +0000 UTC Normal Pod my-jaeger-es-rollover-28351571-rcsdp.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:02 +0000 UTC Normal Pod my-jaeger-es-rollover-28351571-rcsdp.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:05 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28351571 Completed Job completed job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:05 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28351571, status: Complete cronjob-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:05 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28351571 Completed Job completed job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:05 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28351571, status: Complete cronjob-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:06 +0000 UTC Normal Job.batch 02-report-span Completed Job completed job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:07 +0000 UTC Normal Pod 02-check-indices-8d28x Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/02-check-indices-8d28x to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:07 +0000 UTC Normal Pod 02-check-indices-8d28x AddedInterface Add eth0 [10.129.2.39/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:07 +0000 UTC Normal Pod 02-check-indices-8d28x.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:07 +0000 UTC Normal Pod 02-check-indices-8d28x.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:07 +0000 UTC Normal Pod 02-check-indices-8d28x.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:07 +0000 UTC Normal Job.batch 02-check-indices SuccessfulCreate Created pod: 02-check-indices-8d28x job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:10 +0000 UTC Normal Job.batch 02-check-indices Completed Job completed job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:10 +0000 UTC Normal Pod 03-check-indices-rcdtc Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/03-check-indices-rcdtc to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:10 +0000 UTC Normal Pod 03-check-indices-rcdtc AddedInterface Add eth0 [10.129.2.40/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:10 +0000 UTC Normal Pod 03-check-indices-rcdtc.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:10 +0000 UTC Normal Pod 03-check-indices-rcdtc.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:10 +0000 UTC Normal Pod 03-check-indices-rcdtc.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:10 +0000 UTC Normal Job.batch 03-check-indices SuccessfulCreate Created pod: 03-check-indices-rcdtc job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:13 +0000 UTC Normal Job.batch 03-check-indices Completed Job completed job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:13 +0000 UTC Normal Pod 04-check-indices-gk6sx Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/04-check-indices-gk6sx to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:13 +0000 UTC Normal Pod 04-check-indices-gk6sx AddedInterface Add eth0 [10.129.2.41/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:13 +0000 UTC Normal Pod 04-check-indices-gk6sx.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:13 +0000 UTC Normal Pod 04-check-indices-gk6sx.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:13 +0000 UTC Normal Pod 04-check-indices-gk6sx.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:13 +0000 UTC Normal Job.batch 04-check-indices SuccessfulCreate Created pod: 04-check-indices-gk6sx job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:16 +0000 UTC Normal Job.batch 04-check-indices Completed Job completed job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:24 +0000 UTC Normal Pod 03-report-span-vstss Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/03-report-span-vstss to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:24 +0000 UTC Normal Job.batch 03-report-span SuccessfulCreate Created pod: 03-report-span-vstss job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:25 +0000 UTC Normal Pod 03-report-span-vstss AddedInterface Add eth0 [10.129.2.42/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:25 +0000 UTC Normal Pod 03-report-span-vstss.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:25 +0000 UTC Normal Pod 03-report-span-vstss.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:25 +0000 UTC Normal Pod 03-report-span-vstss.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:31 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-7794fb6d5c-c5nql horizontal-pod-autoscaler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:11:48 +0000 UTC Normal Job.batch 03-report-span Completed Job completed job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28351572-kj7sf Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/my-jaeger-es-lookback-28351572-kj7sf to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28351572-kj7sf AddedInterface Add eth0 [10.129.2.44/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28351572-kj7sf.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:edda53eae462f604654d8adc71622101d2f62c9de6ecce4484523341b972723a" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28351572-kj7sf.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28351572-kj7sf.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28351572 SuccessfulCreate Created pod: my-jaeger-es-lookback-28351572-kj7sf job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28351572 cronjob-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28351572-crqws Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/my-jaeger-es-rollover-28351572-crqws to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28351572-crqws AddedInterface Add eth0 [10.129.2.43/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28351572-crqws.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:edda53eae462f604654d8adc71622101d2f62c9de6ecce4484523341b972723a" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28351572-crqws.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28351572-crqws.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28351572 SuccessfulCreate Created pod: my-jaeger-es-rollover-28351572-crqws job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28351572 cronjob-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28351572 Completed Job completed job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28351572, status: Complete cronjob-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28351572 Completed Job completed job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28351572, status: Complete cronjob-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:10 +0000 UTC Normal Pod 05-check-indices-8wdp6 Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/05-check-indices-8wdp6 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:10 +0000 UTC Normal Job.batch 05-check-indices SuccessfulCreate Created pod: 05-check-indices-8wdp6 job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:11 +0000 UTC Normal Pod 05-check-indices-8wdp6 AddedInterface Add eth0 [10.129.2.45/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:11 +0000 UTC Normal Pod 05-check-indices-8wdp6.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:11 +0000 UTC Normal Pod 05-check-indices-8wdp6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:11 +0000 UTC Normal Pod 05-check-indices-8wdp6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:14 +0000 UTC Normal Job.batch 05-check-indices Completed Job completed job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:14 +0000 UTC Normal Pod 06-check-indices-qwzll Binding Scheduled Successfully assigned kuttl-test-capital-shepherd/06-check-indices-qwzll to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:14 +0000 UTC Normal Job.batch 06-check-indices SuccessfulCreate Created pod: 06-check-indices-qwzll job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:15 +0000 UTC Normal Pod 06-check-indices-qwzll AddedInterface Add eth0 [10.129.2.46/23] from ovn-kubernetes logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:15 +0000 UTC Normal Pod 06-check-indices-qwzll.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:15 +0000 UTC Normal Pod 06-check-indices-qwzll.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:15 +0000 UTC Normal Pod 06-check-indices-qwzll.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:12:18 | es-rollover-autoprov | 2023-11-27 14:12:18 +0000 UTC Normal Job.batch 06-check-indices Completed Job completed job-controller logger.go:42: 14:12:18 | es-rollover-autoprov | Deleting namespace: kuttl-test-capital-shepherd === CONT kuttl/harness/es-increasing-replicas logger.go:42: 14:12:26 | es-increasing-replicas | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:12:26 | es-increasing-replicas | Ignoring check-es-nodes.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:12:26 | es-increasing-replicas | Ignoring openshift-check-es-nodes.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:12:26 | es-increasing-replicas | Creating namespace: kuttl-test-prompt-cat logger.go:42: 14:12:26 | es-increasing-replicas/1-install | starting test step 1-install logger.go:42: 14:12:26 | es-increasing-replicas/1-install | Jaeger:kuttl-test-prompt-cat/simple-prod created logger.go:42: 14:13:02 | es-increasing-replicas/1-install | test step completed 1-install logger.go:42: 14:13:02 | es-increasing-replicas/2-install | starting test step 2-install logger.go:42: 14:13:02 | es-increasing-replicas/2-install | Jaeger:kuttl-test-prompt-cat/simple-prod updated logger.go:42: 14:13:08 | es-increasing-replicas/2-install | test step completed 2-install logger.go:42: 14:13:08 | es-increasing-replicas/3-smoke-test | starting test step 3-smoke-test logger.go:42: 14:13:08 | es-increasing-replicas/3-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 14:13:10 | es-increasing-replicas/3-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:13:16 | es-increasing-replicas/3-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 14:13:16 | es-increasing-replicas/3-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:13:16 | es-increasing-replicas/3-smoke-test | job.batch/report-span created logger.go:42: 14:13:16 | es-increasing-replicas/3-smoke-test | job.batch/check-span created logger.go:42: 14:13:29 | es-increasing-replicas/3-smoke-test | test step completed 3-smoke-test logger.go:42: 14:13:29 | es-increasing-replicas/4-install | starting test step 4-install logger.go:42: 14:13:29 | es-increasing-replicas/4-install | Jaeger:kuttl-test-prompt-cat/simple-prod updated logger.go:42: 14:13:29 | es-increasing-replicas/4-install | test step completed 4-install logger.go:42: 14:13:29 | es-increasing-replicas/5-check-es-nodes | starting test step 5-check-es-nodes logger.go:42: 14:13:29 | es-increasing-replicas/5-check-es-nodes | running command: [sh -c ./check-es-nodes.sh $NAMESPACE] logger.go:42: 14:13:29 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 14:13:29 | es-increasing-replicas/5-check-es-nodes | false logger.go:42: 14:13:29 | es-increasing-replicas/5-check-es-nodes | Error: no matches found logger.go:42: 14:13:34 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 14:13:34 | es-increasing-replicas/5-check-es-nodes | true logger.go:42: 14:13:34 | es-increasing-replicas/5-check-es-nodes | test step completed 5-check-es-nodes logger.go:42: 14:13:34 | es-increasing-replicas | es-increasing-replicas events from ns kuttl-test-prompt-cat: logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptcatsimpleprod-1-74869fb9878qzv Binding Scheduled Successfully assigned kuttl-test-prompt-cat/elasticsearch-cdm-kuttltestpromptcatsimpleprod-1-74869fb9878qzv to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptcatsimpleprod-1-74869fb9878qzv AddedInterface Add eth0 [10.128.2.19/23] from ovn-kubernetes logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptcatsimpleprod-1-74869fb9878qzv.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:56ea62bfb0ca36e19a7b21aff3676e49511f05f72da5e76d6427fd8240f328a8" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptcatsimpleprod-1-74869fb9878qzv.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptcatsimpleprod-1-74869fb9878qzv.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptcatsimpleprod-1-74869fb9878qzv.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:79427bea6b5c37894b9782c3821d8b9074838e606daa4a743b2ae060856fa98a" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptcatsimpleprod-1-74869fb9878qzv.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptcatsimpleprod-1-74869fb9878qzv.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:32 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestpromptcatsimpleprod-1-74869fb98b SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestpromptcatsimpleprod-1-74869fb9878qzv replicaset-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:32 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestpromptcatsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestpromptcatsimpleprod-1-74869fb98b to 1 deployment-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:42 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpromptcatsimpleprod-1-74869fb9878qzv.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:47 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpromptcatsimpleprod-1-74869fb9878qzv.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:58 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-77fcbdc546 to 1 deployment-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:58 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-77dffb57c7 to 1 deployment-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:59 +0000 UTC Normal Pod simple-prod-collector-77fcbdc546-cqp4s Binding Scheduled Successfully assigned kuttl-test-prompt-cat/simple-prod-collector-77fcbdc546-cqp4s to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:59 +0000 UTC Normal Pod simple-prod-collector-77fcbdc546-cqp4s AddedInterface Add eth0 [10.131.0.34/23] from ovn-kubernetes logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:59 +0000 UTC Normal Pod simple-prod-collector-77fcbdc546-cqp4s.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:59 +0000 UTC Normal Pod simple-prod-collector-77fcbdc546-cqp4s.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:59 +0000 UTC Normal Pod simple-prod-collector-77fcbdc546-cqp4s.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:59 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-77fcbdc546 SuccessfulCreate Created pod: simple-prod-collector-77fcbdc546-cqp4s replicaset-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:59 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-vvj5b Binding Scheduled Successfully assigned kuttl-test-prompt-cat/simple-prod-query-77dffb57c7-vvj5b to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:59 +0000 UTC Warning Pod simple-prod-query-77dffb57c7-vvj5b FailedMount MountVolume.SetUp failed for volume "simple-prod-ui-oauth-proxy-tls" : secret "simple-prod-ui-oauth-proxy-tls" not found kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:59 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-vvj5b AddedInterface Add eth0 [10.129.2.47/23] from ovn-kubernetes logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:12:59 +0000 UTC Normal ReplicaSet.apps simple-prod-query-77dffb57c7 SuccessfulCreate Created pod: simple-prod-query-77dffb57c7-vvj5b replicaset-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:00 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-vvj5b.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:00 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-vvj5b.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:00 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-vvj5b.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:00 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-vvj5b.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:00 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-vvj5b.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:00 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-vvj5b.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:00 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-vvj5b.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:00 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-vvj5b.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:00 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-vvj5b.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:05 +0000 UTC Normal Pod simple-prod-collector-77fcbdc546-7gs7p Binding Scheduled Successfully assigned kuttl-test-prompt-cat/simple-prod-collector-77fcbdc546-7gs7p to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:05 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-77fcbdc546 SuccessfulCreate Created pod: simple-prod-collector-77fcbdc546-7gs7p replicaset-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:05 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-77fcbdc546 to 2 from 1 deployment-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:05 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-dwzm6 Binding Scheduled Successfully assigned kuttl-test-prompt-cat/simple-prod-query-77dffb57c7-dwzm6 to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:05 +0000 UTC Normal ReplicaSet.apps simple-prod-query-77dffb57c7 SuccessfulCreate Created pod: simple-prod-query-77dffb57c7-dwzm6 replicaset-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:05 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-77dffb57c7 to 2 from 1 deployment-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:06 +0000 UTC Normal Pod simple-prod-collector-77fcbdc546-7gs7p AddedInterface Add eth0 [10.129.2.48/23] from ovn-kubernetes logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:06 +0000 UTC Normal Pod simple-prod-collector-77fcbdc546-7gs7p.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:06 +0000 UTC Normal Pod simple-prod-collector-77fcbdc546-7gs7p.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:06 +0000 UTC Normal Pod simple-prod-collector-77fcbdc546-7gs7p.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:06 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-dwzm6 AddedInterface Add eth0 [10.131.0.35/23] from ovn-kubernetes logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:06 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-dwzm6.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:06 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-dwzm6.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:06 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-dwzm6.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:06 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-dwzm6.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:06 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-dwzm6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:06 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-dwzm6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:06 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-dwzm6.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:06 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-dwzm6.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:06 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-dwzm6.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:11 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-dwzm6.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:11 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-dwzm6.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:11 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-dwzm6.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:11 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-vvj5b.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:11 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-vvj5b.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:11 +0000 UTC Normal Pod simple-prod-query-77dffb57c7-vvj5b.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:11 +0000 UTC Normal ReplicaSet.apps simple-prod-query-77dffb57c7 SuccessfulDelete Deleted pod: simple-prod-query-77dffb57c7-vvj5b replicaset-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:11 +0000 UTC Normal ReplicaSet.apps simple-prod-query-77dffb57c7 SuccessfulDelete Deleted pod: simple-prod-query-77dffb57c7-dwzm6 replicaset-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:11 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-77dffb57c7 to 0 from 2 deployment-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-498cr Binding Scheduled Successfully assigned kuttl-test-prompt-cat/simple-prod-query-854d6cc9c8-498cr to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-498cr AddedInterface Add eth0 [10.129.2.49/23] from ovn-kubernetes logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-498cr.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-498cr.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-498cr.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-498cr.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-498cr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-498cr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-498cr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-qnvsl Binding Scheduled Successfully assigned kuttl-test-prompt-cat/simple-prod-query-854d6cc9c8-qnvsl to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-qnvsl AddedInterface Add eth0 [10.131.0.36/23] from ovn-kubernetes logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-qnvsl.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-qnvsl.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-qnvsl.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-qnvsl.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-qnvsl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-qnvsl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-qnvsl.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal ReplicaSet.apps simple-prod-query-854d6cc9c8 SuccessfulCreate Created pod: simple-prod-query-854d6cc9c8-498cr replicaset-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal ReplicaSet.apps simple-prod-query-854d6cc9c8 SuccessfulCreate Created pod: simple-prod-query-854d6cc9c8-qnvsl replicaset-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:12 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-854d6cc9c8 to 2 deployment-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:13 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-498cr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:13 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-498cr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:13 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-qnvsl.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:13 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-qnvsl.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:16 +0000 UTC Normal Pod report-span-dls8s Binding Scheduled Successfully assigned kuttl-test-prompt-cat/report-span-dls8s to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:16 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-dls8s job-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:17 +0000 UTC Normal Pod check-span-s7l4q Binding Scheduled Successfully assigned kuttl-test-prompt-cat/check-span-s7l4q to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:17 +0000 UTC Normal Pod check-span-s7l4q AddedInterface Add eth0 [10.131.0.37/23] from ovn-kubernetes logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:17 +0000 UTC Normal Pod check-span-s7l4q.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:17 +0000 UTC Normal Pod check-span-s7l4q.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:17 +0000 UTC Normal Pod check-span-s7l4q.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:17 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-s7l4q job-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:17 +0000 UTC Normal Pod report-span-dls8s AddedInterface Add eth0 [10.129.2.50/23] from ovn-kubernetes logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:17 +0000 UTC Normal Pod report-span-dls8s.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:17 +0000 UTC Normal Pod report-span-dls8s.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:17 +0000 UTC Normal Pod report-span-dls8s.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:28 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:30 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptcatsimpleprod-2-84d5f66d845952 Binding Scheduled Successfully assigned kuttl-test-prompt-cat/elasticsearch-cdm-kuttltestpromptcatsimpleprod-2-84d5f66d845952 to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:30 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestpromptcatsimpleprod-2-84d5f66d86 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestpromptcatsimpleprod-2-84d5f66d845952 replicaset-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:30 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestpromptcatsimpleprod-2 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestpromptcatsimpleprod-2-84d5f66d86 to 1 deployment-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:30 +0000 UTC Normal Pod simple-prod-collector-77fcbdc546-7gs7p.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:30 +0000 UTC Normal Pod simple-prod-collector-77fcbdc546-cqp4s.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:30 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-77fcbdc546 SuccessfulDelete Deleted pod: simple-prod-collector-77fcbdc546-cqp4s replicaset-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:30 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-77fcbdc546 SuccessfulDelete Deleted pod: simple-prod-collector-77fcbdc546-7gs7p replicaset-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:30 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled down replica set simple-prod-collector-77fcbdc546 to 0 from 2 deployment-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:30 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-498cr.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:30 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-498cr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:30 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-498cr.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:30 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-qnvsl.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:30 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-qnvsl.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:30 +0000 UTC Normal Pod simple-prod-query-854d6cc9c8-qnvsl.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:30 +0000 UTC Normal ReplicaSet.apps simple-prod-query-854d6cc9c8 SuccessfulDelete Deleted pod: simple-prod-query-854d6cc9c8-498cr replicaset-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:30 +0000 UTC Normal ReplicaSet.apps simple-prod-query-854d6cc9c8 SuccessfulDelete Deleted pod: simple-prod-query-854d6cc9c8-qnvsl replicaset-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:30 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-854d6cc9c8 to 0 from 2 deployment-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:30 +0000 UTC Warning Endpoints simple-prod-query FailedToUpdateEndpoint Failed to update endpoint kuttl-test-prompt-cat/simple-prod-query: Operation cannot be fulfilled on endpoints "simple-prod-query": the object has been modified; please apply your changes to the latest version and try again endpoint-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptcatsimpleprod-2-84d5f66d845952 AddedInterface Add eth0 [10.131.0.38/23] from ovn-kubernetes logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptcatsimpleprod-2-84d5f66d845952.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:56ea62bfb0ca36e19a7b21aff3676e49511f05f72da5e76d6427fd8240f328a8" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptcatsimpleprod-2-84d5f66d845952.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptcatsimpleprod-2-84d5f66d845952.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptcatsimpleprod-2-84d5f66d845952.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:79427bea6b5c37894b9782c3821d8b9074838e606daa4a743b2ae060856fa98a" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptcatsimpleprod-2-84d5f66d845952.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptcatsimpleprod-2-84d5f66d845952.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod simple-prod-collector-5db88495b5-6q7nj Binding Scheduled Successfully assigned kuttl-test-prompt-cat/simple-prod-collector-5db88495b5-6q7nj to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod simple-prod-collector-5db88495b5-fq92q Binding Scheduled Successfully assigned kuttl-test-prompt-cat/simple-prod-collector-5db88495b5-fq92q to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod simple-prod-collector-5db88495b5-fq92q AddedInterface Add eth0 [10.129.2.52/23] from ovn-kubernetes logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod simple-prod-collector-5db88495b5-fq92q.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod simple-prod-collector-5db88495b5-fq92q.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod simple-prod-collector-5db88495b5-fq92q.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5db88495b5 SuccessfulCreate Created pod: simple-prod-collector-5db88495b5-fq92q replicaset-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5db88495b5 SuccessfulCreate Created pod: simple-prod-collector-5db88495b5-6q7nj replicaset-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-5db88495b5 to 2 deployment-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod simple-prod-query-5769549989-kj2g9 Binding Scheduled Successfully assigned kuttl-test-prompt-cat/simple-prod-query-5769549989-kj2g9 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod simple-prod-query-5769549989-kj2g9 AddedInterface Add eth0 [10.129.2.51/23] from ovn-kubernetes logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod simple-prod-query-5769549989-kj2g9.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod simple-prod-query-5769549989-kj2g9.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod simple-prod-query-5769549989-kj2g9.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod simple-prod-query-5769549989-kj2g9.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Pod simple-prod-query-5769549989-zpsl9 Binding Scheduled Successfully assigned kuttl-test-prompt-cat/simple-prod-query-5769549989-zpsl9 to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5769549989 SuccessfulCreate Created pod: simple-prod-query-5769549989-kj2g9 replicaset-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5769549989 SuccessfulCreate Created pod: simple-prod-query-5769549989-zpsl9 replicaset-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:31 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-5769549989 to 2 deployment-controller logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-collector-5db88495b5-6q7nj AddedInterface Add eth0 [10.131.0.40/23] from ovn-kubernetes logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-collector-5db88495b5-6q7nj.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-collector-5db88495b5-6q7nj.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-collector-5db88495b5-6q7nj.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-query-5769549989-kj2g9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-query-5769549989-kj2g9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-query-5769549989-kj2g9.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-query-5769549989-kj2g9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-query-5769549989-kj2g9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-query-5769549989-zpsl9 AddedInterface Add eth0 [10.131.0.39/23] from ovn-kubernetes logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-query-5769549989-zpsl9.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-query-5769549989-zpsl9.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-query-5769549989-zpsl9.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-query-5769549989-zpsl9.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-query-5769549989-zpsl9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-query-5769549989-zpsl9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-query-5769549989-zpsl9.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-query-5769549989-zpsl9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:13:34 | es-increasing-replicas | 2023-11-27 14:13:32 +0000 UTC Normal Pod simple-prod-query-5769549989-zpsl9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:13:34 | es-increasing-replicas | Deleting namespace: kuttl-test-prompt-cat === CONT kuttl/harness/es-index-cleaner-autoprov logger.go:42: 14:14:08 | es-index-cleaner-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:14:08 | es-index-cleaner-autoprov | Creating namespace: kuttl-test-sweeping-thrush logger.go:42: 14:14:08 | es-index-cleaner-autoprov/1-install | starting test step 1-install logger.go:42: 14:14:08 | es-index-cleaner-autoprov/1-install | Jaeger:kuttl-test-sweeping-thrush/test-es-index-cleaner-with-prefix created logger.go:42: 14:14:45 | es-index-cleaner-autoprov/1-install | test step completed 1-install logger.go:42: 14:14:45 | es-index-cleaner-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 14:14:45 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE test-es-index-cleaner-with-prefix /dev/null] logger.go:42: 14:14:46 | es-index-cleaner-autoprov/2-report-spans | Warning: resource jaegers/test-es-index-cleaner-with-prefix is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:14:52 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c DAYS=5 ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 14:14:53 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 14:14:53 | es-index-cleaner-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 14:15:32 | es-index-cleaner-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 14:15:32 | es-index-cleaner-autoprov/3-install | starting test step 3-install logger.go:42: 14:15:32 | es-index-cleaner-autoprov/3-install | Jaeger:kuttl-test-sweeping-thrush/test-es-index-cleaner-with-prefix updated logger.go:42: 14:15:32 | es-index-cleaner-autoprov/3-install | test step completed 3-install logger.go:42: 14:15:32 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | starting test step 4-wait-es-index-cleaner logger.go:42: 14:15:32 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob test-es-index-cleaner-with-prefix-es-index-cleaner --namespace $NAMESPACE] logger.go:42: 14:15:33 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-11-27T14:15:33Z" level=debug msg="Checking if the test-es-index-cleaner-with-prefix-es-index-cleaner CronJob exists" logger.go:42: 14:15:33 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-11-27T14:15:33Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 14:15:33 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-11-27T14:15:33Z" level=warning msg="The BatchV1/Cronjob test-es-index-cleaner-with-prefix-es-index-cleaner was not found" logger.go:42: 14:15:33 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-11-27T14:15:33Z" level=debug msg="Found BatchV/Cronjobs:" logger.go:42: 14:15:33 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-11-27T14:15:33Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 14:15:33 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-11-27T14:15:33Z" level=warning msg="The BatchV1/Cronjob test-es-index-cleaner-with-prefix-es-index-cleaner was not found" logger.go:42: 14:15:33 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-11-27T14:15:33Z" level=debug msg="Found BatchV/Cronjobs:" logger.go:42: 14:15:43 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-11-27T14:15:43Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 14:15:43 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-11-27T14:15:43Z" level=info msg="Cronjob test-es-index-cleaner-with-prefix-es-index-cleaner found successfully" logger.go:42: 14:15:43 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-11-27T14:15:43Z" level=debug msg="Waiting for the next scheduled job from test-es-index-cleaner-with-prefix-es-index-cleaner cronjob" logger.go:42: 14:15:43 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-11-27T14:15:43Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 14:15:43 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-11-27T14:15:43Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 14:15:53 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-11-27T14:15:53Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 14:16:03 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-11-27T14:16:03Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 14:16:13 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2023-11-27T14:16:13Z" level=info msg="Job of owner test-es-index-cleaner-with-prefix-es-index-cleaner succeeded after test-es-index-cleaner-with-prefix-es-index-cleaner 30.022907652s" logger.go:42: 14:16:13 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | test step completed 4-wait-es-index-cleaner logger.go:42: 14:16:13 | es-index-cleaner-autoprov/5-install | starting test step 5-install logger.go:42: 14:16:13 | es-index-cleaner-autoprov/5-install | Jaeger:kuttl-test-sweeping-thrush/test-es-index-cleaner-with-prefix updated logger.go:42: 14:16:13 | es-index-cleaner-autoprov/5-install | test step completed 5-install logger.go:42: 14:16:13 | es-index-cleaner-autoprov/6-check-indices | starting test step 6-check-indices logger.go:42: 14:16:13 | es-index-cleaner-autoprov/6-check-indices | Job:kuttl-test-sweeping-thrush/00-check-indices created logger.go:42: 14:16:16 | es-index-cleaner-autoprov/6-check-indices | test step completed 6-check-indices logger.go:42: 14:16:16 | es-index-cleaner-autoprov | es-index-cleaner-autoprov events from ns kuttl-test-sweeping-thrush: logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:15 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestsweepingthrushtestesindexcl-1-8459b984d6 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestsweepingthrushtestesindexcl-1-8jv785 replicaset-controller logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsweepingthrushtestesindexcl-1-8jv785 Binding Scheduled Successfully assigned kuttl-test-sweeping-thrush/elasticsearch-cdm-kuttltestsweepingthrushtestesindexcl-1-8jv785 to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsweepingthrushtestesindexcl-1-8jv785 AddedInterface Add eth0 [10.128.2.20/23] from ovn-kubernetes logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsweepingthrushtestesindexcl-1-8jv785.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:56ea62bfb0ca36e19a7b21aff3676e49511f05f72da5e76d6427fd8240f328a8" already present on machine kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsweepingthrushtestesindexcl-1-8jv785.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsweepingthrushtestesindexcl-1-8jv785.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsweepingthrushtestesindexcl-1-8jv785.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:79427bea6b5c37894b9782c3821d8b9074838e606daa4a743b2ae060856fa98a" already present on machine kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsweepingthrushtestesindexcl-1-8jv785.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsweepingthrushtestesindexcl-1-8jv785.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:15 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestsweepingthrushtestesindexcl-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestsweepingthrushtestesindexcl-1-8459b984d6 to 1 deployment-controller logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:25 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsweepingthrushtestesindexcl-1-8jv785.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:30 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsweepingthrushtestesindexcl-1-8jv785.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:41 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-7f88446db8-p6pff Binding Scheduled Successfully assigned kuttl-test-sweeping-thrush/test-es-index-cleaner-with-prefix-collector-7f88446db8-p6pff to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:41 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-collector-7f88446db8 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-collector-7f88446db8-p6pff replicaset-controller logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:41 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-collector ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-collector-7f88446db8 to 1 deployment-controller logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:41 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5cdf48fc96-kmsq5 Binding Scheduled Successfully assigned kuttl-test-sweeping-thrush/test-es-index-cleaner-with-prefix-query-5cdf48fc96-kmsq5 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:41 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-5cdf48fc96 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-5cdf48fc96-kmsq5 replicaset-controller logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:41 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-5cdf48fc96 to 1 deployment-controller logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:42 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-7f88446db8-p6pff AddedInterface Add eth0 [10.129.2.53/23] from ovn-kubernetes logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:42 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-7f88446db8-p6pff.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" already present on machine kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:42 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-7f88446db8-p6pff.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:42 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-7f88446db8-p6pff.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:42 +0000 UTC Warning Pod test-es-index-cleaner-with-prefix-query-5cdf48fc96-kmsq5 FailedMount MountVolume.SetUp failed for volume "test-es-index-cleaner-with-prefix-ui-oauth-proxy-tls" : secret "test-es-index-cleaner-with-prefix-ui-oauth-proxy-tls" not found kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:42 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5cdf48fc96-kmsq5 AddedInterface Add eth0 [10.129.2.54/23] from ovn-kubernetes logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:42 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5cdf48fc96-kmsq5.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5cdf48fc96-kmsq5.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5cdf48fc96-kmsq5.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5cdf48fc96-kmsq5.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5cdf48fc96-kmsq5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5cdf48fc96-kmsq5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5cdf48fc96-kmsq5.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5cdf48fc96-kmsq5.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:43 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5cdf48fc96-kmsq5.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:47 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5cdf48fc96-kmsq5.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:47 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5cdf48fc96-kmsq5.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:47 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5cdf48fc96-kmsq5.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:47 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-5cdf48fc96 SuccessfulDelete Deleted pod: test-es-index-cleaner-with-prefix-query-5cdf48fc96-kmsq5 replicaset-controller logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:47 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled down replica set test-es-index-cleaner-with-prefix-query-5cdf48fc96 to 0 from 1 deployment-controller logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:48 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b6b69c8d5-2qqht Binding Scheduled Successfully assigned kuttl-test-sweeping-thrush/test-es-index-cleaner-with-prefix-query-5b6b69c8d5-2qqht to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:48 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-5b6b69c8d5 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-5b6b69c8d5-2qqht replicaset-controller logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:48 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-5b6b69c8d5 to 1 deployment-controller logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b6b69c8d5-2qqht AddedInterface Add eth0 [10.131.0.41/23] from ovn-kubernetes logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b6b69c8d5-2qqht.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b6b69c8d5-2qqht.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b6b69c8d5-2qqht.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b6b69c8d5-2qqht.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b6b69c8d5-2qqht.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b6b69c8d5-2qqht.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b6b69c8d5-2qqht.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b6b69c8d5-2qqht.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:49 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5b6b69c8d5-2qqht.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:53 +0000 UTC Normal Pod 00-report-span-vtbxv Binding Scheduled Successfully assigned kuttl-test-sweeping-thrush/00-report-span-vtbxv to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:53 +0000 UTC Normal Pod 00-report-span-vtbxv AddedInterface Add eth0 [10.129.2.55/23] from ovn-kubernetes logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:53 +0000 UTC Normal Pod 00-report-span-vtbxv.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:53 +0000 UTC Normal Pod 00-report-span-vtbxv.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:53 +0000 UTC Normal Pod 00-report-span-vtbxv.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:53 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-vtbxv job-controller logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:57 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:57 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:14:57 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:15:27 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:15:27 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-7f88446db8-p6pff horizontal-pod-autoscaler logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:15:27 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:15:31 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:16:00 +0000 UTC Normal Job.batch test-es-index-cleaner-with-prefix-es-index-cleaner-28351576 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-es-index-cleaner-2835157r6w98 job-controller logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:16:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2835157r6w98 Binding Scheduled Successfully assigned kuttl-test-sweeping-thrush/test-es-index-cleaner-with-prefix-es-index-cleaner-2835157r6w98 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:16:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2835157r6w98 AddedInterface Add eth0 [10.129.2.57/23] from ovn-kubernetes logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:16:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2835157r6w98.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:08ca2463363916637592e6c1cc1731784e07860269292b216db3e6fd0eb44382" kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:16:00 +0000 UTC Normal CronJob.batch test-es-index-cleaner-with-prefix-es-index-cleaner SuccessfulCreate Created job test-es-index-cleaner-with-prefix-es-index-cleaner-28351576 cronjob-controller logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:16:05 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2835157r6w98.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:08ca2463363916637592e6c1cc1731784e07860269292b216db3e6fd0eb44382" in 4.572s (4.572s including waiting) kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:16:05 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2835157r6w98.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Created Created container test-es-index-cleaner-with-prefix-es-index-cleaner kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:16:05 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2835157r6w98.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Started Started container test-es-index-cleaner-with-prefix-es-index-cleaner kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:16:07 +0000 UTC Normal Job.batch test-es-index-cleaner-with-prefix-es-index-cleaner-28351576 Completed Job completed job-controller logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:16:07 +0000 UTC Normal CronJob.batch test-es-index-cleaner-with-prefix-es-index-cleaner SawCompletedJob Saw completed job: test-es-index-cleaner-with-prefix-es-index-cleaner-28351576, status: Complete cronjob-controller logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:16:13 +0000 UTC Normal Pod 00-check-indices-q4sjj Binding Scheduled Successfully assigned kuttl-test-sweeping-thrush/00-check-indices-q4sjj to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:16:13 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-q4sjj job-controller logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:16:14 +0000 UTC Normal Pod 00-check-indices-q4sjj AddedInterface Add eth0 [10.129.2.58/23] from ovn-kubernetes logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:16:14 +0000 UTC Normal Pod 00-check-indices-q4sjj.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:16:14 +0000 UTC Normal Pod 00-check-indices-q4sjj.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:16:14 +0000 UTC Normal Pod 00-check-indices-q4sjj.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:16:16 | es-index-cleaner-autoprov | 2023-11-27 14:16:16 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 14:16:17 | es-index-cleaner-autoprov | Deleting namespace: kuttl-test-sweeping-thrush === CONT kuttl/harness/es-from-aio-to-production logger.go:42: 14:16:23 | es-from-aio-to-production | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:16:23 | es-from-aio-to-production | Creating namespace: kuttl-test-caring-boa logger.go:42: 14:16:23 | es-from-aio-to-production/0-install | starting test step 0-install logger.go:42: 14:16:23 | es-from-aio-to-production/0-install | Jaeger:kuttl-test-caring-boa/my-jaeger created logger.go:42: 14:16:34 | es-from-aio-to-production/0-install | test step completed 0-install logger.go:42: 14:16:34 | es-from-aio-to-production/1-smoke-test | starting test step 1-smoke-test logger.go:42: 14:16:34 | es-from-aio-to-production/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 14:16:36 | es-from-aio-to-production/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:16:42 | es-from-aio-to-production/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 14:16:43 | es-from-aio-to-production/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:16:43 | es-from-aio-to-production/1-smoke-test | job.batch/report-span created logger.go:42: 14:16:43 | es-from-aio-to-production/1-smoke-test | job.batch/check-span created logger.go:42: 14:16:55 | es-from-aio-to-production/1-smoke-test | test step completed 1-smoke-test logger.go:42: 14:16:55 | es-from-aio-to-production/3-install | starting test step 3-install logger.go:42: 14:16:55 | es-from-aio-to-production/3-install | Jaeger:kuttl-test-caring-boa/my-jaeger updated logger.go:42: 14:17:28 | es-from-aio-to-production/3-install | test step completed 3-install logger.go:42: 14:17:28 | es-from-aio-to-production/4-smoke-test | starting test step 4-smoke-test logger.go:42: 14:17:28 | es-from-aio-to-production/4-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 14:17:35 | es-from-aio-to-production/4-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 14:17:36 | es-from-aio-to-production/4-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:17:36 | es-from-aio-to-production/4-smoke-test | job.batch/report-span unchanged logger.go:42: 14:17:36 | es-from-aio-to-production/4-smoke-test | job.batch/check-span unchanged logger.go:42: 14:17:36 | es-from-aio-to-production/4-smoke-test | test step completed 4-smoke-test logger.go:42: 14:17:36 | es-from-aio-to-production | es-from-aio-to-production events from ns kuttl-test-caring-boa: logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:26 +0000 UTC Normal Pod my-jaeger-764cf88c79-tt5v7 Binding Scheduled Successfully assigned kuttl-test-caring-boa/my-jaeger-764cf88c79-tt5v7 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:26 +0000 UTC Normal ReplicaSet.apps my-jaeger-764cf88c79 SuccessfulCreate Created pod: my-jaeger-764cf88c79-tt5v7 replicaset-controller logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:26 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-764cf88c79 to 1 deployment-controller logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:27 +0000 UTC Warning Pod my-jaeger-764cf88c79-tt5v7 FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-oauth-proxy-tls" : secret "my-jaeger-ui-oauth-proxy-tls" not found kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:27 +0000 UTC Normal Pod my-jaeger-764cf88c79-tt5v7 AddedInterface Add eth0 [10.129.2.59/23] from ovn-kubernetes logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:28 +0000 UTC Normal Pod my-jaeger-764cf88c79-tt5v7.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:32 +0000 UTC Normal Pod my-jaeger-764cf88c79-tt5v7.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" in 4.678s (4.678s including waiting) kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:32 +0000 UTC Normal Pod my-jaeger-764cf88c79-tt5v7.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:32 +0000 UTC Normal Pod my-jaeger-764cf88c79-tt5v7.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:32 +0000 UTC Normal Pod my-jaeger-764cf88c79-tt5v7.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:32 +0000 UTC Normal Pod my-jaeger-764cf88c79-tt5v7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:32 +0000 UTC Normal Pod my-jaeger-764cf88c79-tt5v7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:37 +0000 UTC Normal Pod my-jaeger-764cf88c79-tt5v7.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:37 +0000 UTC Normal Pod my-jaeger-764cf88c79-tt5v7.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:37 +0000 UTC Warning Pod my-jaeger-764cf88c79-tt5v7.spec.containers{jaeger} Unhealthy Readiness probe failed: Get "http://10.129.2.59:14269/": dial tcp 10.129.2.59:14269: connect: connection refused kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:37 +0000 UTC Normal ReplicaSet.apps my-jaeger-764cf88c79 SuccessfulDelete Deleted pod: my-jaeger-764cf88c79-tt5v7 replicaset-controller logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:37 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-764cf88c79 to 0 from 1 deployment-controller logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:38 +0000 UTC Normal Pod my-jaeger-59cd9f9495-6xxdz Binding Scheduled Successfully assigned kuttl-test-caring-boa/my-jaeger-59cd9f9495-6xxdz to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:38 +0000 UTC Normal ReplicaSet.apps my-jaeger-59cd9f9495 SuccessfulCreate Created pod: my-jaeger-59cd9f9495-6xxdz replicaset-controller logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:38 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-59cd9f9495 to 1 deployment-controller logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:39 +0000 UTC Normal Pod my-jaeger-59cd9f9495-6xxdz AddedInterface Add eth0 [10.129.2.60/23] from ovn-kubernetes logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:39 +0000 UTC Normal Pod my-jaeger-59cd9f9495-6xxdz.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:39 +0000 UTC Normal Pod my-jaeger-59cd9f9495-6xxdz.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:39 +0000 UTC Normal Pod my-jaeger-59cd9f9495-6xxdz.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:39 +0000 UTC Normal Pod my-jaeger-59cd9f9495-6xxdz.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:39 +0000 UTC Normal Pod my-jaeger-59cd9f9495-6xxdz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:39 +0000 UTC Normal Pod my-jaeger-59cd9f9495-6xxdz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:43 +0000 UTC Normal Pod check-span-kpv45 Binding Scheduled Successfully assigned kuttl-test-caring-boa/check-span-kpv45 to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:43 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-kpv45 job-controller logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:43 +0000 UTC Normal Pod report-span-jnt5v Binding Scheduled Successfully assigned kuttl-test-caring-boa/report-span-jnt5v to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:43 +0000 UTC Normal Pod report-span-jnt5v AddedInterface Add eth0 [10.128.2.21/23] from ovn-kubernetes logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:43 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-jnt5v job-controller logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:44 +0000 UTC Normal Pod check-span-kpv45 AddedInterface Add eth0 [10.131.0.42/23] from ovn-kubernetes logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:44 +0000 UTC Normal Pod check-span-kpv45.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:44 +0000 UTC Normal Pod check-span-kpv45.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:44 +0000 UTC Normal Pod check-span-kpv45.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:44 +0000 UTC Normal Pod report-span-jnt5v.spec.containers{report-span} Pulling Pulling image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:45 +0000 UTC Normal Pod report-span-jnt5v.spec.containers{report-span} Pulled Successfully pulled image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" in 1.953s (1.953s including waiting) kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:46 +0000 UTC Normal Pod report-span-jnt5v.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:46 +0000 UTC Normal Pod report-span-jnt5v.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:54 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:58 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestcaringboamyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestcaringboamyjaeger-1-65c6fd5f5d to 1 deployment-controller logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcaringboamyjaeger-1-65c6fd5f5d-7dnch Binding Scheduled Successfully assigned kuttl-test-caring-boa/elasticsearch-cdm-kuttltestcaringboamyjaeger-1-65c6fd5f5d-7dnch to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcaringboamyjaeger-1-65c6fd5f5d-7dnch AddedInterface Add eth0 [10.128.2.22/23] from ovn-kubernetes logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcaringboamyjaeger-1-65c6fd5f5d-7dnch.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:56ea62bfb0ca36e19a7b21aff3676e49511f05f72da5e76d6427fd8240f328a8" already present on machine kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcaringboamyjaeger-1-65c6fd5f5d-7dnch.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcaringboamyjaeger-1-65c6fd5f5d-7dnch.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcaringboamyjaeger-1-65c6fd5f5d-7dnch.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:79427bea6b5c37894b9782c3821d8b9074838e606daa4a743b2ae060856fa98a" already present on machine kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcaringboamyjaeger-1-65c6fd5f5d-7dnch.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcaringboamyjaeger-1-65c6fd5f5d-7dnch.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:16:59 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestcaringboamyjaeger-1-65c6fd5f5d SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestcaringboamyjaeger-1-65c6fd5f5d-7dnch replicaset-controller logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:09 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcaringboamyjaeger-1-65c6fd5f5d-7dnch.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:14 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcaringboamyjaeger-1-65c6fd5f5d-7dnch.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:18 +0000 UTC Normal Job.batch report-span Completed Job completed job-controller logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:25 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-cs2cd Binding Scheduled Successfully assigned kuttl-test-caring-boa/my-jaeger-collector-5489f5bd9b-cs2cd to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:25 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5489f5bd9b SuccessfulCreate Created pod: my-jaeger-collector-5489f5bd9b-cs2cd replicaset-controller logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:25 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-5489f5bd9b to 1 deployment-controller logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:25 +0000 UTC Normal Pod my-jaeger-query-8f9d745bb-rw7nv Binding Scheduled Successfully assigned kuttl-test-caring-boa/my-jaeger-query-8f9d745bb-rw7nv to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:25 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-8f9d745bb SuccessfulCreate Created pod: my-jaeger-query-8f9d745bb-rw7nv replicaset-controller logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:25 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-8f9d745bb to 1 deployment-controller logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:26 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-cs2cd AddedInterface Add eth0 [10.131.0.43/23] from ovn-kubernetes logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:26 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-cs2cd.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" already present on machine kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:26 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-cs2cd.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:26 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-cs2cd.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:26 +0000 UTC Normal Pod my-jaeger-query-8f9d745bb-rw7nv AddedInterface Add eth0 [10.131.0.44/23] from ovn-kubernetes logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:26 +0000 UTC Normal Pod my-jaeger-query-8f9d745bb-rw7nv.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:26 +0000 UTC Normal Pod my-jaeger-query-8f9d745bb-rw7nv.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:26 +0000 UTC Normal Pod my-jaeger-query-8f9d745bb-rw7nv.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:26 +0000 UTC Normal Pod my-jaeger-query-8f9d745bb-rw7nv.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:26 +0000 UTC Normal Pod my-jaeger-query-8f9d745bb-rw7nv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:26 +0000 UTC Normal Pod my-jaeger-query-8f9d745bb-rw7nv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:26 +0000 UTC Normal Pod my-jaeger-query-8f9d745bb-rw7nv.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:26 +0000 UTC Normal Pod my-jaeger-query-8f9d745bb-rw7nv.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:26 +0000 UTC Normal Pod my-jaeger-query-8f9d745bb-rw7nv.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:27 +0000 UTC Normal Pod my-jaeger-59cd9f9495-6xxdz.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | 2023-11-27 14:17:27 +0000 UTC Normal Pod my-jaeger-59cd9f9495-6xxdz.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:17:36 | es-from-aio-to-production | Deleting namespace: kuttl-test-caring-boa === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (869.73s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.64s) --- PASS: kuttl/harness/es-multiinstance (138.97s) --- PASS: kuttl/harness/es-streaming-autoprovisioned (209.14s) --- PASS: kuttl/harness/es-simple-prod (5.89s) --- PASS: kuttl/harness/es-rollover-autoprov (193.03s) --- PASS: kuttl/harness/es-increasing-replicas (102.68s) --- PASS: kuttl/harness/es-index-cleaner-autoprov (134.60s) --- PASS: kuttl/harness/es-from-aio-to-production (79.72s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name elasticsearch --report --output /logs/artifacts/elasticsearch.xml ./artifacts/kuttl-report.xml time="2023-11-27T14:17:43Z" level=debug msg="Setting a new name for the test suites" time="2023-11-27T14:17:43Z" level=debug msg="Removing 'artifacts' TestCase" time="2023-11-27T14:17:43Z" level=debug msg="normalizing test case names" time="2023-11-27T14:17:43Z" level=debug msg="elasticsearch/artifacts -> elasticsearch_artifacts" time="2023-11-27T14:17:43Z" level=debug msg="elasticsearch/es-multiinstance -> elasticsearch_es_multiinstance" time="2023-11-27T14:17:43Z" level=debug msg="elasticsearch/es-streaming-autoprovisioned -> elasticsearch_es_streaming_autoprovisioned" time="2023-11-27T14:17:43Z" level=debug msg="elasticsearch/es-simple-prod -> elasticsearch_es_simple_prod" time="2023-11-27T14:17:43Z" level=debug msg="elasticsearch/es-rollover-autoprov -> elasticsearch_es_rollover_autoprov" time="2023-11-27T14:17:43Z" level=debug msg="elasticsearch/es-increasing-replicas -> elasticsearch_es_increasing_replicas" time="2023-11-27T14:17:43Z" level=debug msg="elasticsearch/es-index-cleaner-autoprov -> elasticsearch_es_index_cleaner_autoprov" time="2023-11-27T14:17:43Z" level=debug msg="elasticsearch/es-from-aio-to-production -> elasticsearch_es_from_aio_to_production" +--------------------------------------------+--------+ | NAME | RESULT | +--------------------------------------------+--------+ | elasticsearch_artifacts | passed | | elasticsearch_es_multiinstance | passed | | elasticsearch_es_streaming_autoprovisioned | passed | | elasticsearch_es_simple_prod | passed | | elasticsearch_es_rollover_autoprov | passed | | elasticsearch_es_increasing_replicas | passed | | elasticsearch_es_index_cleaner_autoprov | passed | | elasticsearch_es_from_aio_to_production | passed | +--------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh examples false true + '[' 3 -ne 3 ']' + test_suite_name=examples + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/examples.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-examples make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=0.32.0 \ SKIP_KAFKA=false \ VERTX_IMG=jaegertracing/vertx-create-span:operator-e2e-tests \ ./tests/e2e/examples/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-11-25-110147 True False 21m Cluster version is 4.15.0-0.nightly-2023-11-25-110147' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-11-25-110147 True False 21m Cluster version is 4.15.0-0.nightly-2023-11-25-110147' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 0.32.0 ']' ++ version_le 0.32.0 0.25.0 +++ echo 0.32.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 0.32.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/examples/render.sh ++ export SUITE_DIR=./tests/e2e/examples ++ SUITE_DIR=./tests/e2e/examples ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/examples ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test examples-agent-as-daemonset + '[' 1 -ne 1 ']' + test_name=examples-agent-as-daemonset + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-as-daemonset' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-as-daemonset\e[0m' Rendering files for test examples-agent-as-daemonset + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + mkdir -p examples-agent-as-daemonset + cd examples-agent-as-daemonset + example_name=agent-as-daemonset + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + render_install_example agent-as-daemonset 01 + '[' 2 -ne 2 ']' + example_name=agent-as-daemonset + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/agent-as-daemonset.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + JAEGER_NAME=agent-as-daemonset + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=DaemonSet ++ '[' DaemonSet = null ']' ++ echo DaemonSet ++ return 0 + jaeger_strategy=DaemonSet + '[' DaemonSet = DaemonSet ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example agent-as-daemonset 02 + '[' 2 -ne 2 ']' + example_name=agent-as-daemonset + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/agent-as-daemonset.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/agent-as-daemonset.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/agent-as-daemonset.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/agent-as-daemonset.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + jaeger_name=agent-as-daemonset + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test agent-as-daemonset true 02 + '[' 3 -ne 3 ']' + jaeger=agent-as-daemonset + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-agent-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-agent-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-with-priority-class\e[0m' Rendering files for test examples-agent-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-as-daemonset + '[' examples-agent-as-daemonset '!=' _build ']' + cd .. + mkdir -p examples-agent-with-priority-class + cd examples-agent-with-priority-class + example_name=agent-with-priority-class + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + render_install_example agent-with-priority-class 01 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/agent-with-priority-class.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + JAEGER_NAME=agent-as-daemonset + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=DaemonSet ++ '[' DaemonSet = null ']' ++ echo DaemonSet ++ return 0 + jaeger_strategy=DaemonSet + '[' DaemonSet = DaemonSet ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + jaeger_name=agent-as-daemonset + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test agent-as-daemonset true 02 + '[' 3 -ne 3 ']' + jaeger=agent-as-daemonset + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-all-in-one-with-options + '[' 1 -ne 1 ']' + test_name=examples-all-in-one-with-options + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-all-in-one-with-options' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-all-in-one-with-options\e[0m' Rendering files for test examples-all-in-one-with-options + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-with-priority-class + '[' examples-agent-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-all-in-one-with-options + cd examples-all-in-one-with-options + example_name=all-in-one-with-options + render_install_example all-in-one-with-options 00 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/all-in-one-with-options.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + JAEGER_NAME=my-jaeger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.metadata.name="my-jaeger"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i 'del(.spec.allInOne.image)' ./00-install.yaml + render_smoke_test_example all-in-one-with-options 01 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + jaeger_name=my-jaeger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + sed -i s~my-jaeger-query:443~my-jaeger-query:443/jaeger~gi ./01-smoke-test.yaml + '[' false = true ']' + start_test examples-auto-provision-kafka + '[' 1 -ne 1 ']' + test_name=examples-auto-provision-kafka + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-auto-provision-kafka' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-auto-provision-kafka\e[0m' Rendering files for test examples-auto-provision-kafka + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-all-in-one-with-options + '[' examples-all-in-one-with-options '!=' _build ']' + cd .. + mkdir -p examples-auto-provision-kafka + cd examples-auto-provision-kafka + example_name=auto-provision-kafka + render_install_kafka_operator 01 + '[' 1 -ne 1 ']' + test_step=01 + '[' true '!=' true ']' + render_install_example auto-provision-kafka 02 + '[' 2 -ne 2 ']' + example_name=auto-provision-kafka + test_step=02 + install_file=./02-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/auto-provision-kafka.yaml -o ./02-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./02-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./02-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./02-install.yaml ++ jaeger_name=auto-provision-kafka ++ '[' -z auto-provision-kafka ']' ++ echo auto-provision-kafka ++ return 0 + JAEGER_NAME=auto-provision-kafka + local jaeger_strategy ++ get_jaeger_strategy ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./02-install.yaml ++ strategy=streaming ++ '[' streaming = production ']' ++ '[' streaming = streaming ']' ++ echo streaming ++ return 0 + jaeger_strategy=streaming + '[' streaming = DaemonSet ']' + '[' streaming = allInOne ']' + '[' streaming = production ']' + '[' streaming = streaming ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/streaming-jaeger-assert.yaml.template -o ./02-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./02-install.yaml + mv ./02-assert.yaml ./05-assert.yaml + render_assert_kafka true auto-provision-kafka 02 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provision-kafka + test_step=02 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provision-kafka + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./02-assert.yaml ++ expr 02 + 1 + CLUSTER_NAME=auto-provision-kafka + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./3-assert.yaml ++ expr 02 + 2 + CLUSTER_NAME=auto-provision-kafka + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./04-assert.yaml + render_smoke_test_example auto-provision-kafka 06 + '[' 2 -ne 2 ']' + example_name=auto-provision-kafka + test_step=06 + deployment_file=/tmp/jaeger-tests/examples/auto-provision-kafka.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/auto-provision-kafka.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/auto-provision-kafka.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/auto-provision-kafka.yaml ++ jaeger_name=auto-provision-kafka ++ '[' -z auto-provision-kafka ']' ++ echo auto-provision-kafka ++ return 0 + jaeger_name=auto-provision-kafka + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test auto-provision-kafka true 06 + '[' 3 -ne 3 ']' + jaeger=auto-provision-kafka + is_secured=true + test_step=06 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://auto-provision-kafka-query:443 + JAEGER_QUERY_ENDPOINT=https://auto-provision-kafka-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://auto-provision-kafka-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://auto-provision-kafka-collector-headless:14268 + export JAEGER_NAME=auto-provision-kafka + JAEGER_NAME=auto-provision-kafka + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./06-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-business-application-injected-sidecar + '[' 1 -ne 1 ']' + test_name=examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-business-application-injected-sidecar' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-business-application-injected-sidecar\e[0m' Rendering files for test examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-auto-provision-kafka + '[' examples-auto-provision-kafka '!=' _build ']' + cd .. + mkdir -p examples-business-application-injected-sidecar + cd examples-business-application-injected-sidecar + example_name=simplest + cp /tmp/jaeger-tests/examples/business-application-injected-sidecar.yaml ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].image=strenv(VERTX_IMG)' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.path="/"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.port=8080' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.initialDelaySeconds=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.failureThreshold=3' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.periodSeconds=10' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.successThreshold=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.timeoutSeconds=1' ./00-install.yaml + render_install_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example simplest 02 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 02 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-collector-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-collector-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-collector-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-collector-with-priority-class\e[0m' Rendering files for test examples-collector-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-business-application-injected-sidecar + '[' examples-business-application-injected-sidecar '!=' _build ']' + cd .. + mkdir -p examples-collector-with-priority-class + cd examples-collector-with-priority-class + example_name=collector-with-priority-class + render_install_example collector-with-priority-class 00 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/collector-with-priority-class.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + JAEGER_NAME=collector-with-high-priority + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example collector-with-priority-class 01 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + jaeger_name=collector-with-high-priority + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test collector-with-high-priority true 01 + '[' 3 -ne 3 ']' + jaeger=collector-with-high-priority + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + export JAEGER_NAME=collector-with-high-priority + JAEGER_NAME=collector-with-high-priority + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-service-types + '[' 1 -ne 1 ']' + test_name=examples-service-types + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-service-types' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-service-types\e[0m' Rendering files for test examples-service-types + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-collector-with-priority-class + '[' examples-collector-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-service-types + cd examples-service-types + example_name=service-types + render_install_example service-types 00 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/service-types.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + JAEGER_NAME=service-types + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example service-types 01 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/service-types.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/service-types.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/service-types.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/service-types.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + jaeger_name=service-types + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test service-types true 01 + '[' 3 -ne 3 ']' + jaeger=service-types + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + export JAEGER_NAME=service-types + JAEGER_NAME=service-types + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod + '[' 1 -ne 1 ']' + test_name=examples-simple-prod + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod\e[0m' Rendering files for test examples-simple-prod + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-service-types + '[' examples-service-types '!=' _build ']' + cd .. + mkdir -p examples-simple-prod + cd examples-simple-prod + example_name=simple-prod + render_install_example simple-prod 01 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod 02 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod-with-volumes + '[' 1 -ne 1 ']' + test_name=examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod-with-volumes' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod-with-volumes\e[0m' Rendering files for test examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod + '[' examples-simple-prod '!=' _build ']' + cd .. + mkdir -p examples-simple-prod-with-volumes + cd examples-simple-prod-with-volumes + example_name=simple-prod-with-volumes + render_install_example simple-prod-with-volumes 01 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod-with-volumes 02 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + /tmp/jaeger-tests/bin/gomplate -f ./03-check-volume.yaml.template -o 03-check-volume.yaml + start_test examples-simplest + '[' 1 -ne 1 ']' + test_name=examples-simplest + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simplest' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simplest\e[0m' Rendering files for test examples-simplest + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod-with-volumes + '[' examples-simple-prod-with-volumes '!=' _build ']' + cd .. + mkdir -p examples-simplest + cd examples-simplest + example_name=simplest + render_install_example simplest 00 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 01 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger + '[' 1 -ne 1 ']' + test_name=examples-with-badger + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger\e[0m' Rendering files for test examples-with-badger + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simplest + '[' examples-simplest '!=' _build ']' + cd .. + mkdir -p examples-with-badger + cd examples-with-badger + example_name=with-badger + render_install_example with-badger 00 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + JAEGER_NAME=with-badger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger 01 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + jaeger_name=with-badger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + export JAEGER_NAME=with-badger + JAEGER_NAME=with-badger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger-and-volume + '[' 1 -ne 1 ']' + test_name=examples-with-badger-and-volume + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger-and-volume' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger-and-volume\e[0m' Rendering files for test examples-with-badger-and-volume + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger + '[' examples-with-badger '!=' _build ']' + cd .. + mkdir -p examples-with-badger-and-volume + cd examples-with-badger-and-volume + example_name=with-badger-and-volume + render_install_example with-badger-and-volume 00 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger-and-volume.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + JAEGER_NAME=with-badger-and-volume + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger-and-volume 01 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + jaeger_name=with-badger-and-volume + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger-and-volume true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger-and-volume + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + export JAEGER_NAME=with-badger-and-volume + JAEGER_NAME=with-badger-and-volume + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-cassandra + '[' 1 -ne 1 ']' + test_name=examples-with-cassandra + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-cassandra' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-cassandra\e[0m' Rendering files for test examples-with-cassandra + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger-and-volume + '[' examples-with-badger-and-volume '!=' _build ']' + cd .. + mkdir -p examples-with-cassandra + cd examples-with-cassandra + example_name=with-cassandra + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-cassandra 01 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-cassandra.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + JAEGER_NAME=with-cassandra + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-cassandra 02 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-cassandra.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-cassandra.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + jaeger_name=with-cassandra + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-cassandra true 02 + '[' 3 -ne 3 ']' + jaeger=with-cassandra + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + export JAEGER_NAME=with-cassandra + JAEGER_NAME=with-cassandra + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-sampling + '[' 1 -ne 1 ']' + test_name=examples-with-sampling + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-sampling' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-sampling\e[0m' Rendering files for test examples-with-sampling + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-cassandra + '[' examples-with-cassandra '!=' _build ']' + cd .. + mkdir -p examples-with-sampling + cd examples-with-sampling + export example_name=with-sampling + example_name=with-sampling + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-sampling 01 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-sampling.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + JAEGER_NAME=with-sampling + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-sampling 02 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-sampling.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-sampling.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + jaeger_name=with-sampling + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-sampling true 02 + '[' 3 -ne 3 ']' + jaeger=with-sampling + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + export JAEGER_NAME=with-sampling + JAEGER_NAME=with-sampling + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + start_test examples-openshift-agent-as-daemonset + '[' 1 -ne 1 ']' + test_name=examples-openshift-agent-as-daemonset + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-openshift-agent-as-daemonset' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-openshift-agent-as-daemonset\e[0m' Rendering files for test examples-openshift-agent-as-daemonset + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-sampling + '[' examples-with-sampling '!=' _build ']' + cd .. + mkdir -p examples-openshift-agent-as-daemonset + cd examples-openshift-agent-as-daemonset + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/agent-as-daemonset.yaml -o 02-install.yaml + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./02-assert.yaml + render_install_vertx 03 + '[' 1 -ne 1 ']' + test_step=03 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./03-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].env=[{"name": "JAEGER_AGENT_HOST", "valueFrom": {"fieldRef": {"apiVersion": "v1", "fieldPath": "status.hostIP"}}}]' ./03-install.yaml + render_find_service agent-as-daemonset production order 00 04 + '[' 5 -ne 5 ']' + jaeger=agent-as-daemonset + deployment_strategy=production + service_name=order + job_number=00 + test_step=04 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' production '!=' allInOne ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/find-service.yaml.template -o ./04-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + start_test examples-openshift-with-htpasswd + '[' 1 -ne 1 ']' + test_name=examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-openshift-with-htpasswd' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-openshift-with-htpasswd\e[0m' Rendering files for test examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-openshift-agent-as-daemonset + '[' examples-openshift-agent-as-daemonset '!=' _build ']' + cd .. + mkdir -p examples-openshift-with-htpasswd + cd examples-openshift-with-htpasswd + export JAEGER_NAME=with-htpasswd + JAEGER_NAME=with-htpasswd + export JAEGER_USERNAME=awesomeuser + JAEGER_USERNAME=awesomeuser + export JAEGER_PASSWORD=awesomepassword + JAEGER_PASSWORD=awesomepassword + export 'JAEGER_USER_PASSWORD_HASH=awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' + JAEGER_USER_PASSWORD_HASH='awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ echo 'awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ base64 + SECRET=YXdlc29tZXVzZXI6e1NIQX11VWRxUFZVeXFOQm1FUlUwUXhqM0tGYVpuanc9Cg== + /tmp/jaeger-tests/bin/gomplate -f ./00-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/with-htpasswd.yaml -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + export 'GET_URL_COMMAND=kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + export 'URL=https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + INSECURE=true + JAEGER_USERNAME= + JAEGER_PASSWORD= + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-unsecured.yaml + JAEGER_USERNAME=wronguser + JAEGER_PASSWORD=wrongpassword + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-check-unauthorized.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./04-check-authorized.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running examples E2E tests' Running examples E2E tests + cd tests/e2e/examples/_build + set +e + KUBECONFIG=/tmp/kubeconfig-4018005699 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 17 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/examples-agent-as-daemonset === PAUSE kuttl/harness/examples-agent-as-daemonset === RUN kuttl/harness/examples-agent-with-priority-class === PAUSE kuttl/harness/examples-agent-with-priority-class === RUN kuttl/harness/examples-all-in-one-with-options === PAUSE kuttl/harness/examples-all-in-one-with-options === RUN kuttl/harness/examples-auto-provision-kafka === PAUSE kuttl/harness/examples-auto-provision-kafka === RUN kuttl/harness/examples-business-application-injected-sidecar === PAUSE kuttl/harness/examples-business-application-injected-sidecar === RUN kuttl/harness/examples-collector-with-priority-class === PAUSE kuttl/harness/examples-collector-with-priority-class === RUN kuttl/harness/examples-openshift-agent-as-daemonset === PAUSE kuttl/harness/examples-openshift-agent-as-daemonset === RUN kuttl/harness/examples-openshift-with-htpasswd === PAUSE kuttl/harness/examples-openshift-with-htpasswd === RUN kuttl/harness/examples-service-types === PAUSE kuttl/harness/examples-service-types === RUN kuttl/harness/examples-simple-prod === PAUSE kuttl/harness/examples-simple-prod === RUN kuttl/harness/examples-simple-prod-with-volumes === PAUSE kuttl/harness/examples-simple-prod-with-volumes === RUN kuttl/harness/examples-simplest === PAUSE kuttl/harness/examples-simplest === RUN kuttl/harness/examples-with-badger === PAUSE kuttl/harness/examples-with-badger === RUN kuttl/harness/examples-with-badger-and-volume === PAUSE kuttl/harness/examples-with-badger-and-volume === RUN kuttl/harness/examples-with-cassandra === PAUSE kuttl/harness/examples-with-cassandra === RUN kuttl/harness/examples-with-sampling === PAUSE kuttl/harness/examples-with-sampling === CONT kuttl/harness/artifacts logger.go:42: 14:18:25 | artifacts | Creating namespace: kuttl-test-vocal-sheep logger.go:42: 14:18:25 | artifacts | artifacts events from ns kuttl-test-vocal-sheep: logger.go:42: 14:18:25 | artifacts | Deleting namespace: kuttl-test-vocal-sheep === CONT kuttl/harness/examples-service-types logger.go:42: 14:18:30 | examples-service-types | Creating namespace: kuttl-test-big-piglet logger.go:42: 14:18:30 | examples-service-types/0-install | starting test step 0-install logger.go:42: 14:18:30 | examples-service-types/0-install | Jaeger:kuttl-test-big-piglet/service-types created logger.go:42: 14:18:37 | examples-service-types/0-install | test step completed 0-install logger.go:42: 14:18:37 | examples-service-types/1-smoke-test | starting test step 1-smoke-test logger.go:42: 14:18:37 | examples-service-types/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE service-types /dev/null] logger.go:42: 14:18:38 | examples-service-types/1-smoke-test | Warning: resource jaegers/service-types is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:18:44 | examples-service-types/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://service-types-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 14:18:45 | examples-service-types/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:18:45 | examples-service-types/1-smoke-test | job.batch/report-span created logger.go:42: 14:18:45 | examples-service-types/1-smoke-test | job.batch/check-span created logger.go:42: 14:18:57 | examples-service-types/1-smoke-test | test step completed 1-smoke-test logger.go:42: 14:18:57 | examples-service-types/2- | starting test step 2- logger.go:42: 14:18:57 | examples-service-types/2- | test step completed 2- logger.go:42: 14:18:57 | examples-service-types | examples-service-types events from ns kuttl-test-big-piglet: logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:34 +0000 UTC Normal Pod service-types-5958b97566-gfbc5 Binding Scheduled Successfully assigned kuttl-test-big-piglet/service-types-5958b97566-gfbc5 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:34 +0000 UTC Normal ReplicaSet.apps service-types-5958b97566 SuccessfulCreate Created pod: service-types-5958b97566-gfbc5 replicaset-controller logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:34 +0000 UTC Normal Service service-types-collector EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:34 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-5958b97566 to 1 deployment-controller logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:35 +0000 UTC Normal Pod service-types-5958b97566-gfbc5 AddedInterface Add eth0 [10.129.2.61/23] from ovn-kubernetes logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:35 +0000 UTC Normal Pod service-types-5958b97566-gfbc5.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:35 +0000 UTC Normal Pod service-types-5958b97566-gfbc5.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:35 +0000 UTC Normal Pod service-types-5958b97566-gfbc5.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:35 +0000 UTC Normal Pod service-types-5958b97566-gfbc5.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:35 +0000 UTC Normal Pod service-types-5958b97566-gfbc5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:35 +0000 UTC Normal Pod service-types-5958b97566-gfbc5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:37 +0000 UTC Normal Service service-types-collector EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:37 +0000 UTC Normal Service service-types-query EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:40 +0000 UTC Normal Pod service-types-5958b97566-gfbc5.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:40 +0000 UTC Normal Pod service-types-5958b97566-gfbc5.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:40 +0000 UTC Normal ReplicaSet.apps service-types-5958b97566 SuccessfulDelete Deleted pod: service-types-5958b97566-gfbc5 replicaset-controller logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:40 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled down replica set service-types-5958b97566 to 0 from 1 deployment-controller logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:41 +0000 UTC Normal Pod service-types-776d6c86d7-zfzn2 Binding Scheduled Successfully assigned kuttl-test-big-piglet/service-types-776d6c86d7-zfzn2 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:41 +0000 UTC Normal Pod service-types-776d6c86d7-zfzn2 AddedInterface Add eth0 [10.129.2.62/23] from ovn-kubernetes logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:41 +0000 UTC Normal Pod service-types-776d6c86d7-zfzn2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:41 +0000 UTC Normal Pod service-types-776d6c86d7-zfzn2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:41 +0000 UTC Normal Pod service-types-776d6c86d7-zfzn2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:41 +0000 UTC Normal Pod service-types-776d6c86d7-zfzn2.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:41 +0000 UTC Normal Pod service-types-776d6c86d7-zfzn2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:41 +0000 UTC Normal Pod service-types-776d6c86d7-zfzn2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:41 +0000 UTC Normal ReplicaSet.apps service-types-776d6c86d7 SuccessfulCreate Created pod: service-types-776d6c86d7-zfzn2 replicaset-controller logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:41 +0000 UTC Normal Service service-types-query EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:41 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-776d6c86d7 to 1 deployment-controller logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:45 +0000 UTC Normal Pod check-span-xdcsb Binding Scheduled Successfully assigned kuttl-test-big-piglet/check-span-xdcsb to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:45 +0000 UTC Normal Pod check-span-xdcsb AddedInterface Add eth0 [10.131.0.45/23] from ovn-kubernetes logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:45 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-xdcsb job-controller logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:45 +0000 UTC Normal Pod report-span-r9h8w Binding Scheduled Successfully assigned kuttl-test-big-piglet/report-span-r9h8w to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:45 +0000 UTC Normal Pod report-span-r9h8w AddedInterface Add eth0 [10.128.2.23/23] from ovn-kubernetes logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:45 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-r9h8w job-controller logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:46 +0000 UTC Normal Pod check-span-xdcsb.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:46 +0000 UTC Normal Pod check-span-xdcsb.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:46 +0000 UTC Normal Pod check-span-xdcsb.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:46 +0000 UTC Normal Pod report-span-r9h8w.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:46 +0000 UTC Normal Pod report-span-r9h8w.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:46 +0000 UTC Normal Pod report-span-r9h8w.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:18:57 | examples-service-types | 2023-11-27 14:18:57 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:18:57 | examples-service-types | Deleting namespace: kuttl-test-big-piglet === CONT kuttl/harness/examples-with-sampling logger.go:42: 14:19:24 | examples-with-sampling | Creating namespace: kuttl-test-current-hyena logger.go:42: 14:19:24 | examples-with-sampling/0-install | starting test step 0-install logger.go:42: 14:19:24 | examples-with-sampling/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 14:19:24 | examples-with-sampling/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 14:19:24 | examples-with-sampling/0-install | >>>> Creating namespace kuttl-test-current-hyena logger.go:42: 14:19:24 | examples-with-sampling/0-install | kubectl create namespace kuttl-test-current-hyena 2>&1 | grep -v "already exists" || true logger.go:42: 14:19:24 | examples-with-sampling/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-current-hyena 2>&1 | grep -v "already exists" || true logger.go:42: 14:19:25 | examples-with-sampling/0-install | service/cassandra created logger.go:42: 14:19:25 | examples-with-sampling/0-install | statefulset.apps/cassandra created logger.go:42: 14:19:25 | examples-with-sampling/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 14:19:36 | examples-with-sampling/0-install | test step completed 0-install logger.go:42: 14:19:36 | examples-with-sampling/1-install | starting test step 1-install logger.go:42: 14:19:36 | examples-with-sampling/1-install | Jaeger:kuttl-test-current-hyena/with-sampling created logger.go:42: 14:19:46 | examples-with-sampling/1-install | test step completed 1-install logger.go:42: 14:19:46 | examples-with-sampling/2-smoke-test | starting test step 2-smoke-test logger.go:42: 14:19:46 | examples-with-sampling/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-sampling /dev/null] logger.go:42: 14:19:47 | examples-with-sampling/2-smoke-test | Warning: resource jaegers/with-sampling is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:19:54 | examples-with-sampling/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 14:19:54 | examples-with-sampling/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:19:55 | examples-with-sampling/2-smoke-test | job.batch/report-span created logger.go:42: 14:19:55 | examples-with-sampling/2-smoke-test | job.batch/check-span created logger.go:42: 14:20:07 | examples-with-sampling/2-smoke-test | test step completed 2-smoke-test logger.go:42: 14:20:07 | examples-with-sampling/3- | starting test step 3- logger.go:42: 14:20:07 | examples-with-sampling/3- | test step completed 3- logger.go:42: 14:20:07 | examples-with-sampling | examples-with-sampling events from ns kuttl-test-current-hyena: logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:25 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-current-hyena/cassandra-0 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:25 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.129.2.63/23] from ovn-kubernetes logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:25 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:25 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:29 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 3.872s (3.872s including waiting) kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:29 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:29 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:30 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-current-hyena/cassandra-1 to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:30 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:31 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.131.0.46/23] from ovn-kubernetes logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:31 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:35 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 3.943s (3.943s including waiting) kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:35 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:35 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:39 +0000 UTC Normal Pod with-sampling-656c6ddbf8-s5qlp Binding Scheduled Successfully assigned kuttl-test-current-hyena/with-sampling-656c6ddbf8-s5qlp to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:39 +0000 UTC Normal ReplicaSet.apps with-sampling-656c6ddbf8 SuccessfulCreate Created pod: with-sampling-656c6ddbf8-s5qlp replicaset-controller logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:39 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-656c6ddbf8 to 1 deployment-controller logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:40 +0000 UTC Normal Pod with-sampling-656c6ddbf8-s5qlp AddedInterface Add eth0 [10.128.2.24/23] from ovn-kubernetes logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:40 +0000 UTC Normal Pod with-sampling-656c6ddbf8-s5qlp.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:44 +0000 UTC Normal Pod with-sampling-656c6ddbf8-s5qlp.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" in 4.244s (4.244s including waiting) kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:44 +0000 UTC Normal Pod with-sampling-656c6ddbf8-s5qlp.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:44 +0000 UTC Normal Pod with-sampling-656c6ddbf8-s5qlp.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:44 +0000 UTC Normal Pod with-sampling-656c6ddbf8-s5qlp.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:44 +0000 UTC Normal Pod with-sampling-656c6ddbf8-s5qlp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:44 +0000 UTC Normal Pod with-sampling-656c6ddbf8-s5qlp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:50 +0000 UTC Normal Pod with-sampling-656c6ddbf8-s5qlp.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:50 +0000 UTC Normal Pod with-sampling-656c6ddbf8-s5qlp.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:50 +0000 UTC Normal ReplicaSet.apps with-sampling-656c6ddbf8 SuccessfulDelete Deleted pod: with-sampling-656c6ddbf8-s5qlp replicaset-controller logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:50 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled down replica set with-sampling-656c6ddbf8 to 0 from 1 deployment-controller logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:51 +0000 UTC Normal Pod with-sampling-6cbf6979bb-cb2fc Binding Scheduled Successfully assigned kuttl-test-current-hyena/with-sampling-6cbf6979bb-cb2fc to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:51 +0000 UTC Normal Pod with-sampling-6cbf6979bb-cb2fc AddedInterface Add eth0 [10.129.2.64/23] from ovn-kubernetes logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:51 +0000 UTC Normal Pod with-sampling-6cbf6979bb-cb2fc.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:51 +0000 UTC Normal Pod with-sampling-6cbf6979bb-cb2fc.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:51 +0000 UTC Normal Pod with-sampling-6cbf6979bb-cb2fc.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:51 +0000 UTC Normal Pod with-sampling-6cbf6979bb-cb2fc.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:51 +0000 UTC Normal Pod with-sampling-6cbf6979bb-cb2fc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:51 +0000 UTC Normal Pod with-sampling-6cbf6979bb-cb2fc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:51 +0000 UTC Normal ReplicaSet.apps with-sampling-6cbf6979bb SuccessfulCreate Created pod: with-sampling-6cbf6979bb-cb2fc replicaset-controller logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:51 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-6cbf6979bb to 1 deployment-controller logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:55 +0000 UTC Normal Pod check-span-45hhk Binding Scheduled Successfully assigned kuttl-test-current-hyena/check-span-45hhk to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:55 +0000 UTC Normal Pod check-span-45hhk AddedInterface Add eth0 [10.128.2.26/23] from ovn-kubernetes logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:55 +0000 UTC Normal Pod check-span-45hhk.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:55 +0000 UTC Normal Pod check-span-45hhk.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:55 +0000 UTC Normal Pod check-span-45hhk.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:55 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-45hhk job-controller logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:55 +0000 UTC Normal Pod report-span-vxh5k Binding Scheduled Successfully assigned kuttl-test-current-hyena/report-span-vxh5k to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:55 +0000 UTC Normal Pod report-span-vxh5k AddedInterface Add eth0 [10.128.2.25/23] from ovn-kubernetes logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:55 +0000 UTC Normal Pod report-span-vxh5k.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:55 +0000 UTC Normal Pod report-span-vxh5k.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:55 +0000 UTC Normal Pod report-span-vxh5k.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:19:55 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-vxh5k job-controller logger.go:42: 14:20:07 | examples-with-sampling | 2023-11-27 14:20:06 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:20:07 | examples-with-sampling | Deleting namespace: kuttl-test-current-hyena === CONT kuttl/harness/examples-with-cassandra logger.go:42: 14:20:24 | examples-with-cassandra | Creating namespace: kuttl-test-classic-garfish logger.go:42: 14:20:24 | examples-with-cassandra/0-install | starting test step 0-install logger.go:42: 14:20:24 | examples-with-cassandra/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 14:20:24 | examples-with-cassandra/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 14:20:24 | examples-with-cassandra/0-install | >>>> Creating namespace kuttl-test-classic-garfish logger.go:42: 14:20:24 | examples-with-cassandra/0-install | kubectl create namespace kuttl-test-classic-garfish 2>&1 | grep -v "already exists" || true logger.go:42: 14:20:24 | examples-with-cassandra/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-classic-garfish 2>&1 | grep -v "already exists" || true logger.go:42: 14:20:24 | examples-with-cassandra/0-install | service/cassandra created logger.go:42: 14:20:24 | examples-with-cassandra/0-install | statefulset.apps/cassandra created logger.go:42: 14:20:24 | examples-with-cassandra/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 14:20:30 | examples-with-cassandra/0-install | test step completed 0-install logger.go:42: 14:20:30 | examples-with-cassandra/1-install | starting test step 1-install logger.go:42: 14:20:30 | examples-with-cassandra/1-install | Jaeger:kuttl-test-classic-garfish/with-cassandra created logger.go:42: 14:21:03 | examples-with-cassandra/1-install | test step completed 1-install logger.go:42: 14:21:03 | examples-with-cassandra/2-smoke-test | starting test step 2-smoke-test logger.go:42: 14:21:03 | examples-with-cassandra/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-cassandra /dev/null] logger.go:42: 14:21:04 | examples-with-cassandra/2-smoke-test | Warning: resource jaegers/with-cassandra is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:21:10 | examples-with-cassandra/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 14:21:10 | examples-with-cassandra/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:21:11 | examples-with-cassandra/2-smoke-test | job.batch/report-span created logger.go:42: 14:21:11 | examples-with-cassandra/2-smoke-test | job.batch/check-span created logger.go:42: 14:21:22 | examples-with-cassandra/2-smoke-test | test step completed 2-smoke-test logger.go:42: 14:21:22 | examples-with-cassandra | examples-with-cassandra events from ns kuttl-test-classic-garfish: logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:24 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-classic-garfish/cassandra-0 to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:24 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:25 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.128.2.27/23] from ovn-kubernetes logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:25 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:29 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 3.936s (3.936s including waiting) kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:29 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:29 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:29 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-classic-garfish/cassandra-1 to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:29 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.131.0.47/23] from ovn-kubernetes logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:29 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:29 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:29 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:29 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:34 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-8wmmk Binding Scheduled Successfully assigned kuttl-test-classic-garfish/with-cassandra-cassandra-schema-job-8wmmk to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:34 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-8wmmk AddedInterface Add eth0 [10.129.2.65/23] from ovn-kubernetes logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:34 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-8wmmk.spec.containers{with-cassandra-cassandra-schema-job} Pulling Pulling image "jaegertracing/jaeger-cassandra-schema:1.47.0" kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:34 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job SuccessfulCreate Created pod: with-cassandra-cassandra-schema-job-8wmmk job-controller logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:39 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-8wmmk.spec.containers{with-cassandra-cassandra-schema-job} Pulled Successfully pulled image "jaegertracing/jaeger-cassandra-schema:1.47.0" in 4.665s (4.665s including waiting) kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:39 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-8wmmk.spec.containers{with-cassandra-cassandra-schema-job} Created Created container with-cassandra-cassandra-schema-job kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:39 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-8wmmk.spec.containers{with-cassandra-cassandra-schema-job} Started Started container with-cassandra-cassandra-schema-job kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:45 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job Completed Job completed job-controller logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:46 +0000 UTC Normal Pod with-cassandra-5f9687fc48-9vtsp Binding Scheduled Successfully assigned kuttl-test-classic-garfish/with-cassandra-5f9687fc48-9vtsp to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:46 +0000 UTC Normal Pod with-cassandra-5f9687fc48-9vtsp AddedInterface Add eth0 [10.129.2.66/23] from ovn-kubernetes logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:46 +0000 UTC Normal Pod with-cassandra-5f9687fc48-9vtsp.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:46 +0000 UTC Normal Pod with-cassandra-5f9687fc48-9vtsp.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:46 +0000 UTC Normal ReplicaSet.apps with-cassandra-5f9687fc48 SuccessfulCreate Created pod: with-cassandra-5f9687fc48-9vtsp replicaset-controller logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:46 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-5f9687fc48 to 1 deployment-controller logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:47 +0000 UTC Normal Pod with-cassandra-5f9687fc48-9vtsp.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:47 +0000 UTC Normal Pod with-cassandra-5f9687fc48-9vtsp.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:47 +0000 UTC Normal Pod with-cassandra-5f9687fc48-9vtsp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:47 +0000 UTC Normal Pod with-cassandra-5f9687fc48-9vtsp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:20:48 +0000 UTC Warning Pod with-cassandra-5f9687fc48-9vtsp.spec.containers{jaeger} BackOff Back-off restarting failed container jaeger in pod with-cassandra-5f9687fc48-9vtsp_kuttl-test-classic-garfish(13753856-6724-4d6f-b6e4-b5152a57f19c) kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:05 +0000 UTC Normal Pod with-cassandra-5f9687fc48-9vtsp.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:05 +0000 UTC Normal Pod with-cassandra-5f9687fc48-9vtsp.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:05 +0000 UTC Normal ReplicaSet.apps with-cassandra-5f9687fc48 SuccessfulDelete Deleted pod: with-cassandra-5f9687fc48-9vtsp replicaset-controller logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:05 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled down replica set with-cassandra-5f9687fc48 to 0 from 1 deployment-controller logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:06 +0000 UTC Normal Pod with-cassandra-785f74796d-b6lwh Binding Scheduled Successfully assigned kuttl-test-classic-garfish/with-cassandra-785f74796d-b6lwh to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:06 +0000 UTC Normal ReplicaSet.apps with-cassandra-785f74796d SuccessfulCreate Created pod: with-cassandra-785f74796d-b6lwh replicaset-controller logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:06 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-785f74796d to 1 deployment-controller logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:07 +0000 UTC Normal Pod with-cassandra-785f74796d-b6lwh AddedInterface Add eth0 [10.129.2.67/23] from ovn-kubernetes logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:07 +0000 UTC Normal Pod with-cassandra-785f74796d-b6lwh.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:07 +0000 UTC Normal Pod with-cassandra-785f74796d-b6lwh.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:07 +0000 UTC Normal Pod with-cassandra-785f74796d-b6lwh.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:07 +0000 UTC Normal Pod with-cassandra-785f74796d-b6lwh.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:07 +0000 UTC Normal Pod with-cassandra-785f74796d-b6lwh.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:07 +0000 UTC Normal Pod with-cassandra-785f74796d-b6lwh.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:11 +0000 UTC Normal Pod check-span-cg56c Binding Scheduled Successfully assigned kuttl-test-classic-garfish/check-span-cg56c to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:11 +0000 UTC Normal Pod check-span-cg56c AddedInterface Add eth0 [10.128.2.29/23] from ovn-kubernetes logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:11 +0000 UTC Normal Pod check-span-cg56c.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:11 +0000 UTC Normal Pod check-span-cg56c.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:11 +0000 UTC Normal Pod check-span-cg56c.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:11 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-cg56c job-controller logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:11 +0000 UTC Normal Pod report-span-rkq5j Binding Scheduled Successfully assigned kuttl-test-classic-garfish/report-span-rkq5j to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:11 +0000 UTC Normal Pod report-span-rkq5j AddedInterface Add eth0 [10.128.2.28/23] from ovn-kubernetes logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:11 +0000 UTC Normal Pod report-span-rkq5j.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:11 +0000 UTC Normal Pod report-span-rkq5j.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:11 +0000 UTC Normal Pod report-span-rkq5j.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:11 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-rkq5j job-controller logger.go:42: 14:21:22 | examples-with-cassandra | 2023-11-27 14:21:22 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:21:22 | examples-with-cassandra | Deleting namespace: kuttl-test-classic-garfish === CONT kuttl/harness/examples-with-badger-and-volume logger.go:42: 14:21:34 | examples-with-badger-and-volume | Creating namespace: kuttl-test-valued-perch logger.go:42: 14:21:34 | examples-with-badger-and-volume/0-install | starting test step 0-install logger.go:42: 14:21:34 | examples-with-badger-and-volume/0-install | Jaeger:kuttl-test-valued-perch/with-badger-and-volume created logger.go:42: 14:21:40 | examples-with-badger-and-volume/0-install | test step completed 0-install logger.go:42: 14:21:40 | examples-with-badger-and-volume/1-smoke-test | starting test step 1-smoke-test logger.go:42: 14:21:40 | examples-with-badger-and-volume/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger-and-volume /dev/null] logger.go:42: 14:21:41 | examples-with-badger-and-volume/1-smoke-test | Warning: resource jaegers/with-badger-and-volume is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:21:47 | examples-with-badger-and-volume/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 14:21:48 | examples-with-badger-and-volume/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:21:48 | examples-with-badger-and-volume/1-smoke-test | job.batch/report-span created logger.go:42: 14:21:48 | examples-with-badger-and-volume/1-smoke-test | job.batch/check-span created logger.go:42: 14:21:59 | examples-with-badger-and-volume/1-smoke-test | test step completed 1-smoke-test logger.go:42: 14:21:59 | examples-with-badger-and-volume | examples-with-badger-and-volume events from ns kuttl-test-valued-perch: logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:37 +0000 UTC Normal Pod with-badger-and-volume-588b77cc6c-p8gwc Binding Scheduled Successfully assigned kuttl-test-valued-perch/with-badger-and-volume-588b77cc6c-p8gwc to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:37 +0000 UTC Normal ReplicaSet.apps with-badger-and-volume-588b77cc6c SuccessfulCreate Created pod: with-badger-and-volume-588b77cc6c-p8gwc replicaset-controller logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:37 +0000 UTC Normal Deployment.apps with-badger-and-volume ScalingReplicaSet Scaled up replica set with-badger-and-volume-588b77cc6c to 1 deployment-controller logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:38 +0000 UTC Normal Pod with-badger-and-volume-588b77cc6c-p8gwc AddedInterface Add eth0 [10.128.2.30/23] from ovn-kubernetes logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:38 +0000 UTC Normal Pod with-badger-and-volume-588b77cc6c-p8gwc.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:38 +0000 UTC Normal Pod with-badger-and-volume-588b77cc6c-p8gwc.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:38 +0000 UTC Normal Pod with-badger-and-volume-588b77cc6c-p8gwc.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:38 +0000 UTC Normal Pod with-badger-and-volume-588b77cc6c-p8gwc.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:38 +0000 UTC Normal Pod with-badger-and-volume-588b77cc6c-p8gwc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:38 +0000 UTC Normal Pod with-badger-and-volume-588b77cc6c-p8gwc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:44 +0000 UTC Normal Pod with-badger-and-volume-588b77cc6c-p8gwc.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:44 +0000 UTC Normal Pod with-badger-and-volume-588b77cc6c-p8gwc.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:44 +0000 UTC Normal ReplicaSet.apps with-badger-and-volume-588b77cc6c SuccessfulDelete Deleted pod: with-badger-and-volume-588b77cc6c-p8gwc replicaset-controller logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:44 +0000 UTC Normal Deployment.apps with-badger-and-volume ScalingReplicaSet Scaled down replica set with-badger-and-volume-588b77cc6c to 0 from 1 deployment-controller logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:45 +0000 UTC Normal Pod with-badger-and-volume-54d64df564-kfmqp Binding Scheduled Successfully assigned kuttl-test-valued-perch/with-badger-and-volume-54d64df564-kfmqp to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:45 +0000 UTC Normal Pod with-badger-and-volume-54d64df564-kfmqp AddedInterface Add eth0 [10.129.2.68/23] from ovn-kubernetes logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:45 +0000 UTC Normal Pod with-badger-and-volume-54d64df564-kfmqp.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:45 +0000 UTC Normal Pod with-badger-and-volume-54d64df564-kfmqp.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:45 +0000 UTC Normal Pod with-badger-and-volume-54d64df564-kfmqp.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:45 +0000 UTC Normal Pod with-badger-and-volume-54d64df564-kfmqp.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:45 +0000 UTC Normal ReplicaSet.apps with-badger-and-volume-54d64df564 SuccessfulCreate Created pod: with-badger-and-volume-54d64df564-kfmqp replicaset-controller logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:45 +0000 UTC Normal Deployment.apps with-badger-and-volume ScalingReplicaSet Scaled up replica set with-badger-and-volume-54d64df564 to 1 deployment-controller logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:46 +0000 UTC Normal Pod with-badger-and-volume-54d64df564-kfmqp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:46 +0000 UTC Normal Pod with-badger-and-volume-54d64df564-kfmqp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:48 +0000 UTC Normal Pod check-span-qc9l4 Binding Scheduled Successfully assigned kuttl-test-valued-perch/check-span-qc9l4 to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:48 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-qc9l4 job-controller logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:48 +0000 UTC Normal Pod report-span-46bvb Binding Scheduled Successfully assigned kuttl-test-valued-perch/report-span-46bvb to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:48 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-46bvb job-controller logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:49 +0000 UTC Normal Pod check-span-qc9l4 AddedInterface Add eth0 [10.131.0.48/23] from ovn-kubernetes logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:49 +0000 UTC Normal Pod check-span-qc9l4.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:49 +0000 UTC Normal Pod check-span-qc9l4.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:49 +0000 UTC Normal Pod check-span-qc9l4.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:49 +0000 UTC Normal Pod report-span-46bvb AddedInterface Add eth0 [10.128.2.31/23] from ovn-kubernetes logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:49 +0000 UTC Normal Pod report-span-46bvb.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:49 +0000 UTC Normal Pod report-span-46bvb.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:49 +0000 UTC Normal Pod report-span-46bvb.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:21:59 | examples-with-badger-and-volume | 2023-11-27 14:21:59 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:21:59 | examples-with-badger-and-volume | Deleting namespace: kuttl-test-valued-perch === CONT kuttl/harness/examples-with-badger logger.go:42: 14:22:11 | examples-with-badger | Creating namespace: kuttl-test-joint-quetzal logger.go:42: 14:22:11 | examples-with-badger/0-install | starting test step 0-install logger.go:42: 14:22:11 | examples-with-badger/0-install | Jaeger:kuttl-test-joint-quetzal/with-badger created logger.go:42: 14:22:17 | examples-with-badger/0-install | test step completed 0-install logger.go:42: 14:22:17 | examples-with-badger/1-smoke-test | starting test step 1-smoke-test logger.go:42: 14:22:17 | examples-with-badger/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger /dev/null] logger.go:42: 14:22:18 | examples-with-badger/1-smoke-test | Warning: resource jaegers/with-badger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:22:25 | examples-with-badger/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 14:22:25 | examples-with-badger/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:22:25 | examples-with-badger/1-smoke-test | job.batch/report-span created logger.go:42: 14:22:25 | examples-with-badger/1-smoke-test | job.batch/check-span created logger.go:42: 14:22:36 | examples-with-badger/1-smoke-test | test step completed 1-smoke-test logger.go:42: 14:22:36 | examples-with-badger | examples-with-badger events from ns kuttl-test-joint-quetzal: logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:14 +0000 UTC Normal Pod with-badger-558fd89df7-gtq2h Binding Scheduled Successfully assigned kuttl-test-joint-quetzal/with-badger-558fd89df7-gtq2h to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:14 +0000 UTC Normal ReplicaSet.apps with-badger-558fd89df7 SuccessfulCreate Created pod: with-badger-558fd89df7-gtq2h replicaset-controller logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:14 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-558fd89df7 to 1 deployment-controller logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:15 +0000 UTC Normal Pod with-badger-558fd89df7-gtq2h AddedInterface Add eth0 [10.129.2.69/23] from ovn-kubernetes logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:15 +0000 UTC Normal Pod with-badger-558fd89df7-gtq2h.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:15 +0000 UTC Normal Pod with-badger-558fd89df7-gtq2h.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:15 +0000 UTC Normal Pod with-badger-558fd89df7-gtq2h.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:15 +0000 UTC Normal Pod with-badger-558fd89df7-gtq2h.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:15 +0000 UTC Normal Pod with-badger-558fd89df7-gtq2h.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:15 +0000 UTC Normal Pod with-badger-558fd89df7-gtq2h.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:22 +0000 UTC Normal Pod with-badger-558fd89df7-gtq2h.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:22 +0000 UTC Normal Pod with-badger-558fd89df7-gtq2h.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:22 +0000 UTC Normal ReplicaSet.apps with-badger-558fd89df7 SuccessfulDelete Deleted pod: with-badger-558fd89df7-gtq2h replicaset-controller logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:22 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled down replica set with-badger-558fd89df7 to 0 from 1 deployment-controller logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:23 +0000 UTC Normal Pod with-badger-f6b44ccc7-z7lfs Binding Scheduled Successfully assigned kuttl-test-joint-quetzal/with-badger-f6b44ccc7-z7lfs to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:23 +0000 UTC Normal Pod with-badger-f6b44ccc7-z7lfs AddedInterface Add eth0 [10.129.2.70/23] from ovn-kubernetes logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:23 +0000 UTC Normal Pod with-badger-f6b44ccc7-z7lfs.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:23 +0000 UTC Normal Pod with-badger-f6b44ccc7-z7lfs.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:23 +0000 UTC Normal Pod with-badger-f6b44ccc7-z7lfs.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:23 +0000 UTC Normal Pod with-badger-f6b44ccc7-z7lfs.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:23 +0000 UTC Normal Pod with-badger-f6b44ccc7-z7lfs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:23 +0000 UTC Normal Pod with-badger-f6b44ccc7-z7lfs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:23 +0000 UTC Normal ReplicaSet.apps with-badger-f6b44ccc7 SuccessfulCreate Created pod: with-badger-f6b44ccc7-z7lfs replicaset-controller logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:23 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-f6b44ccc7 to 1 deployment-controller logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:25 +0000 UTC Normal Pod check-span-6nqbd Binding Scheduled Successfully assigned kuttl-test-joint-quetzal/check-span-6nqbd to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:25 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-6nqbd job-controller logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:25 +0000 UTC Normal Pod report-span-p8k84 Binding Scheduled Successfully assigned kuttl-test-joint-quetzal/report-span-p8k84 to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:25 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-p8k84 job-controller logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:26 +0000 UTC Normal Pod check-span-6nqbd AddedInterface Add eth0 [10.131.0.49/23] from ovn-kubernetes logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:26 +0000 UTC Normal Pod check-span-6nqbd.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:26 +0000 UTC Normal Pod check-span-6nqbd.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:26 +0000 UTC Normal Pod check-span-6nqbd.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:26 +0000 UTC Normal Pod report-span-p8k84 AddedInterface Add eth0 [10.128.2.32/23] from ovn-kubernetes logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:26 +0000 UTC Normal Pod report-span-p8k84.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:26 +0000 UTC Normal Pod report-span-p8k84.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:26 +0000 UTC Normal Pod report-span-p8k84.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:22:36 | examples-with-badger | 2023-11-27 14:22:36 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:22:36 | examples-with-badger | Deleting namespace: kuttl-test-joint-quetzal === CONT kuttl/harness/examples-simplest logger.go:42: 14:22:48 | examples-simplest | Creating namespace: kuttl-test-quality-grubworm logger.go:42: 14:22:48 | examples-simplest/0-install | starting test step 0-install logger.go:42: 14:22:48 | examples-simplest/0-install | Jaeger:kuttl-test-quality-grubworm/simplest created logger.go:42: 14:22:55 | examples-simplest/0-install | test step completed 0-install logger.go:42: 14:22:55 | examples-simplest/1-smoke-test | starting test step 1-smoke-test logger.go:42: 14:22:55 | examples-simplest/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 14:22:56 | examples-simplest/1-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:23:03 | examples-simplest/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 14:23:03 | examples-simplest/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:23:03 | examples-simplest/1-smoke-test | job.batch/report-span created logger.go:42: 14:23:03 | examples-simplest/1-smoke-test | job.batch/check-span created logger.go:42: 14:23:14 | examples-simplest/1-smoke-test | test step completed 1-smoke-test logger.go:42: 14:23:14 | examples-simplest | examples-simplest events from ns kuttl-test-quality-grubworm: logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:52 +0000 UTC Normal Pod simplest-689c64b5f4-q5c9l Binding Scheduled Successfully assigned kuttl-test-quality-grubworm/simplest-689c64b5f4-q5c9l to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:52 +0000 UTC Normal Pod simplest-689c64b5f4-q5c9l AddedInterface Add eth0 [10.128.2.33/23] from ovn-kubernetes logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:52 +0000 UTC Normal Pod simplest-689c64b5f4-q5c9l.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:52 +0000 UTC Normal Pod simplest-689c64b5f4-q5c9l.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:52 +0000 UTC Normal Pod simplest-689c64b5f4-q5c9l.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:52 +0000 UTC Normal Pod simplest-689c64b5f4-q5c9l.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:52 +0000 UTC Normal Pod simplest-689c64b5f4-q5c9l.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:52 +0000 UTC Normal Pod simplest-689c64b5f4-q5c9l.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:52 +0000 UTC Normal ReplicaSet.apps simplest-689c64b5f4 SuccessfulCreate Created pod: simplest-689c64b5f4-q5c9l replicaset-controller logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:52 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-689c64b5f4 to 1 deployment-controller logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:57 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-689c64b5f4 to 0 from 1 deployment-controller logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:58 +0000 UTC Normal Pod simplest-57df67b6cb-jvlt5 Binding Scheduled Successfully assigned kuttl-test-quality-grubworm/simplest-57df67b6cb-jvlt5 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:58 +0000 UTC Normal ReplicaSet.apps simplest-57df67b6cb SuccessfulCreate Created pod: simplest-57df67b6cb-jvlt5 replicaset-controller logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:58 +0000 UTC Normal Pod simplest-689c64b5f4-q5c9l.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:58 +0000 UTC Normal Pod simplest-689c64b5f4-q5c9l.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:58 +0000 UTC Normal ReplicaSet.apps simplest-689c64b5f4 SuccessfulDelete Deleted pod: simplest-689c64b5f4-q5c9l replicaset-controller logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:58 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-57df67b6cb to 1 deployment-controller logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:59 +0000 UTC Normal Pod simplest-57df67b6cb-jvlt5 AddedInterface Add eth0 [10.129.2.71/23] from ovn-kubernetes logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:59 +0000 UTC Normal Pod simplest-57df67b6cb-jvlt5.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:59 +0000 UTC Normal Pod simplest-57df67b6cb-jvlt5.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:59 +0000 UTC Normal Pod simplest-57df67b6cb-jvlt5.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:59 +0000 UTC Normal Pod simplest-57df67b6cb-jvlt5.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:59 +0000 UTC Normal Pod simplest-57df67b6cb-jvlt5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:22:59 +0000 UTC Normal Pod simplest-57df67b6cb-jvlt5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:23:03 +0000 UTC Normal Pod check-span-mtj84 Binding Scheduled Successfully assigned kuttl-test-quality-grubworm/check-span-mtj84 to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:23:03 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-mtj84 job-controller logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:23:03 +0000 UTC Normal Pod report-span-5h7hv Binding Scheduled Successfully assigned kuttl-test-quality-grubworm/report-span-5h7hv to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:23:03 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-5h7hv job-controller logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:23:04 +0000 UTC Normal Pod check-span-mtj84 AddedInterface Add eth0 [10.131.0.50/23] from ovn-kubernetes logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:23:04 +0000 UTC Normal Pod check-span-mtj84.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:23:04 +0000 UTC Normal Pod check-span-mtj84.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:23:04 +0000 UTC Normal Pod check-span-mtj84.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:23:04 +0000 UTC Normal Pod report-span-5h7hv AddedInterface Add eth0 [10.128.2.34/23] from ovn-kubernetes logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:23:04 +0000 UTC Normal Pod report-span-5h7hv.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:23:04 +0000 UTC Normal Pod report-span-5h7hv.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:23:04 +0000 UTC Normal Pod report-span-5h7hv.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:23:14 | examples-simplest | 2023-11-27 14:23:14 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:23:14 | examples-simplest | Deleting namespace: kuttl-test-quality-grubworm === CONT kuttl/harness/examples-simple-prod-with-volumes logger.go:42: 14:23:26 | examples-simple-prod-with-volumes | Ignoring 03-check-volume.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:23:26 | examples-simple-prod-with-volumes | Creating namespace: kuttl-test-tops-mackerel logger.go:42: 14:23:26 | examples-simple-prod-with-volumes/1-install | starting test step 1-install logger.go:42: 14:23:26 | examples-simple-prod-with-volumes/1-install | Jaeger:kuttl-test-tops-mackerel/simple-prod created logger.go:42: 14:24:01 | examples-simple-prod-with-volumes/1-install | test step completed 1-install logger.go:42: 14:24:01 | examples-simple-prod-with-volumes/2-smoke-test | starting test step 2-smoke-test logger.go:42: 14:24:01 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 14:24:03 | examples-simple-prod-with-volumes/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:24:09 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 14:24:09 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:24:09 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/report-span created logger.go:42: 14:24:09 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/check-span created logger.go:42: 14:24:22 | examples-simple-prod-with-volumes/2-smoke-test | test step completed 2-smoke-test logger.go:42: 14:24:22 | examples-simple-prod-with-volumes/3-check-volume | starting test step 3-check-volume logger.go:42: 14:24:22 | examples-simple-prod-with-volumes/3-check-volume | running command: [sh -c kubectl exec $(kubectl get pods -n $NAMESPACE -l app=jaeger -l app.kubernetes.io/component=collector -o yaml | /tmp/jaeger-tests/bin/yq e '.items[0].metadata.name') -n $NAMESPACE -- ls /usr/share/elasticsearch/data] logger.go:42: 14:24:22 | examples-simple-prod-with-volumes/3-check-volume | test step completed 3-check-volume logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | examples-simple-prod-with-volumes events from ns kuttl-test-tops-mackerel: logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopsmackerelsimpleprod-1-99c8dd4qwsj Binding Scheduled Successfully assigned kuttl-test-tops-mackerel/elasticsearch-cdm-kuttltesttopsmackerelsimpleprod-1-99c8dd4qwsj to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopsmackerelsimpleprod-1-99c8dd4qwsj AddedInterface Add eth0 [10.128.2.35/23] from ovn-kubernetes logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopsmackerelsimpleprod-1-99c8dd4qwsj.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:56ea62bfb0ca36e19a7b21aff3676e49511f05f72da5e76d6427fd8240f328a8" already present on machine kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopsmackerelsimpleprod-1-99c8dd4qwsj.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopsmackerelsimpleprod-1-99c8dd4qwsj.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopsmackerelsimpleprod-1-99c8dd4qwsj.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:79427bea6b5c37894b9782c3821d8b9074838e606daa4a743b2ae060856fa98a" already present on machine kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopsmackerelsimpleprod-1-99c8dd4qwsj.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopsmackerelsimpleprod-1-99c8dd4qwsj.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:32 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltesttopsmackerelsimpleprod-1-99c8dd8c4 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltesttopsmackerelsimpleprod-1-99c8dd4qwsj replicaset-controller logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:32 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltesttopsmackerelsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltesttopsmackerelsimpleprod-1-99c8dd8c4 to 1 deployment-controller logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:42 +0000 UTC Warning Pod elasticsearch-cdm-kuttltesttopsmackerelsimpleprod-1-99c8dd4qwsj.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:47 +0000 UTC Warning Pod elasticsearch-cdm-kuttltesttopsmackerelsimpleprod-1-99c8dd4qwsj.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal Pod simple-prod-collector-55ff468b9d-47sx8 Binding Scheduled Successfully assigned kuttl-test-tops-mackerel/simple-prod-collector-55ff468b9d-47sx8 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal Pod simple-prod-collector-55ff468b9d-47sx8 AddedInterface Add eth0 [10.129.2.72/23] from ovn-kubernetes logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal Pod simple-prod-collector-55ff468b9d-47sx8.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" already present on machine kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal Pod simple-prod-collector-55ff468b9d-47sx8.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal Pod simple-prod-collector-55ff468b9d-47sx8.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-55ff468b9d SuccessfulCreate Created pod: simple-prod-collector-55ff468b9d-47sx8 replicaset-controller logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-55ff468b9d to 1 deployment-controller logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal Pod simple-prod-query-64d4f8f49f-6hdvr Binding Scheduled Successfully assigned kuttl-test-tops-mackerel/simple-prod-query-64d4f8f49f-6hdvr to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal Pod simple-prod-query-64d4f8f49f-6hdvr AddedInterface Add eth0 [10.129.2.73/23] from ovn-kubernetes logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal Pod simple-prod-query-64d4f8f49f-6hdvr.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal Pod simple-prod-query-64d4f8f49f-6hdvr.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal Pod simple-prod-query-64d4f8f49f-6hdvr.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal Pod simple-prod-query-64d4f8f49f-6hdvr.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal Pod simple-prod-query-64d4f8f49f-6hdvr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal Pod simple-prod-query-64d4f8f49f-6hdvr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal Pod simple-prod-query-64d4f8f49f-6hdvr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal Pod simple-prod-query-64d4f8f49f-6hdvr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal Pod simple-prod-query-64d4f8f49f-6hdvr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal ReplicaSet.apps simple-prod-query-64d4f8f49f SuccessfulCreate Created pod: simple-prod-query-64d4f8f49f-6hdvr replicaset-controller logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:23:59 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-64d4f8f49f to 1 deployment-controller logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:04 +0000 UTC Normal Pod simple-prod-query-64d4f8f49f-6hdvr.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:04 +0000 UTC Normal Pod simple-prod-query-64d4f8f49f-6hdvr.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:04 +0000 UTC Normal Pod simple-prod-query-64d4f8f49f-6hdvr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:04 +0000 UTC Normal ReplicaSet.apps simple-prod-query-64d4f8f49f SuccessfulDelete Deleted pod: simple-prod-query-64d4f8f49f-6hdvr replicaset-controller logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:04 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-64d4f8f49f to 0 from 1 deployment-controller logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:05 +0000 UTC Normal Pod simple-prod-query-6f4ddd649b-z2ld9 Binding Scheduled Successfully assigned kuttl-test-tops-mackerel/simple-prod-query-6f4ddd649b-z2ld9 to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:05 +0000 UTC Normal Pod simple-prod-query-6f4ddd649b-z2ld9 AddedInterface Add eth0 [10.131.0.51/23] from ovn-kubernetes logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:05 +0000 UTC Normal Pod simple-prod-query-6f4ddd649b-z2ld9.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:05 +0000 UTC Normal Pod simple-prod-query-6f4ddd649b-z2ld9.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:05 +0000 UTC Normal Pod simple-prod-query-6f4ddd649b-z2ld9.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:05 +0000 UTC Normal Pod simple-prod-query-6f4ddd649b-z2ld9.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:05 +0000 UTC Normal Pod simple-prod-query-6f4ddd649b-z2ld9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:05 +0000 UTC Normal Pod simple-prod-query-6f4ddd649b-z2ld9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:05 +0000 UTC Normal Pod simple-prod-query-6f4ddd649b-z2ld9.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:05 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6f4ddd649b SuccessfulCreate Created pod: simple-prod-query-6f4ddd649b-z2ld9 replicaset-controller logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:05 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6f4ddd649b to 1 deployment-controller logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:06 +0000 UTC Normal Pod simple-prod-query-6f4ddd649b-z2ld9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:06 +0000 UTC Normal Pod simple-prod-query-6f4ddd649b-z2ld9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:09 +0000 UTC Normal Pod report-span-snz8p Binding Scheduled Successfully assigned kuttl-test-tops-mackerel/report-span-snz8p to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:09 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-snz8p job-controller logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:10 +0000 UTC Normal Pod check-span-h4tt6 Binding Scheduled Successfully assigned kuttl-test-tops-mackerel/check-span-h4tt6 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:10 +0000 UTC Normal Pod check-span-h4tt6 AddedInterface Add eth0 [10.129.2.75/23] from ovn-kubernetes logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:10 +0000 UTC Normal Pod check-span-h4tt6.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:10 +0000 UTC Normal Pod check-span-h4tt6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:10 +0000 UTC Normal Pod check-span-h4tt6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:10 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-h4tt6 job-controller logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:10 +0000 UTC Normal Pod report-span-snz8p AddedInterface Add eth0 [10.129.2.74/23] from ovn-kubernetes logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:10 +0000 UTC Normal Pod report-span-snz8p.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:10 +0000 UTC Normal Pod report-span-snz8p.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:10 +0000 UTC Normal Pod report-span-snz8p.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | 2023-11-27 14:24:21 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:24:22 | examples-simple-prod-with-volumes | Deleting namespace: kuttl-test-tops-mackerel === CONT kuttl/harness/examples-simple-prod logger.go:42: 14:24:34 | examples-simple-prod | Creating namespace: kuttl-test-top-wahoo logger.go:42: 14:24:34 | examples-simple-prod/1-install | starting test step 1-install logger.go:42: 14:24:34 | examples-simple-prod/1-install | Jaeger:kuttl-test-top-wahoo/simple-prod created logger.go:42: 14:25:11 | examples-simple-prod/1-install | test step completed 1-install logger.go:42: 14:25:11 | examples-simple-prod/2-smoke-test | starting test step 2-smoke-test logger.go:42: 14:25:11 | examples-simple-prod/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 14:25:12 | examples-simple-prod/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:25:19 | examples-simple-prod/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 14:25:19 | examples-simple-prod/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:25:19 | examples-simple-prod/2-smoke-test | job.batch/report-span created logger.go:42: 14:25:19 | examples-simple-prod/2-smoke-test | job.batch/check-span created logger.go:42: 14:25:32 | examples-simple-prod/2-smoke-test | test step completed 2-smoke-test logger.go:42: 14:25:32 | examples-simple-prod | examples-simple-prod events from ns kuttl-test-top-wahoo: logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:24:41 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltesttopwahoosimpleprod-1-58d79f78dd SuccessfulCreate Created pod: elasticsearch-cdm-kuttltesttopwahoosimpleprod-1-58d79f78ddkjl2z replicaset-controller logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:24:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopwahoosimpleprod-1-58d79f78ddkjl2z Binding Scheduled Successfully assigned kuttl-test-top-wahoo/elasticsearch-cdm-kuttltesttopwahoosimpleprod-1-58d79f78ddkjl2z to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:24:41 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltesttopwahoosimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltesttopwahoosimpleprod-1-58d79f78dd to 1 deployment-controller logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:24:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopwahoosimpleprod-1-58d79f78ddkjl2z AddedInterface Add eth0 [10.128.2.36/23] from ovn-kubernetes logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:24:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopwahoosimpleprod-1-58d79f78ddkjl2z.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:56ea62bfb0ca36e19a7b21aff3676e49511f05f72da5e76d6427fd8240f328a8" already present on machine kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:24:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopwahoosimpleprod-1-58d79f78ddkjl2z.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:24:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopwahoosimpleprod-1-58d79f78ddkjl2z.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:24:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopwahoosimpleprod-1-58d79f78ddkjl2z.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:79427bea6b5c37894b9782c3821d8b9074838e606daa4a743b2ae060856fa98a" already present on machine kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:24:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopwahoosimpleprod-1-58d79f78ddkjl2z.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:24:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttopwahoosimpleprod-1-58d79f78ddkjl2z.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:24:52 +0000 UTC Warning Pod elasticsearch-cdm-kuttltesttopwahoosimpleprod-1-58d79f78ddkjl2z.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:24:57 +0000 UTC Warning Pod elasticsearch-cdm-kuttltesttopwahoosimpleprod-1-58d79f78ddkjl2z.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:08 +0000 UTC Normal Pod simple-prod-collector-77fcbdc546-mn6fm Binding Scheduled Successfully assigned kuttl-test-top-wahoo/simple-prod-collector-77fcbdc546-mn6fm to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:08 +0000 UTC Warning Pod simple-prod-collector-77fcbdc546-mn6fm FailedMount MountVolume.SetUp failed for volume "simple-prod-collector-tls-config-volume" : secret "simple-prod-collector-headless-tls" not found kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:08 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-77fcbdc546 SuccessfulCreate Created pod: simple-prod-collector-77fcbdc546-mn6fm replicaset-controller logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:08 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-77fcbdc546 to 1 deployment-controller logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:08 +0000 UTC Normal Pod simple-prod-query-6455f7d7f7-75pxj Binding Scheduled Successfully assigned kuttl-test-top-wahoo/simple-prod-query-6455f7d7f7-75pxj to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:08 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6455f7d7f7 SuccessfulCreate Created pod: simple-prod-query-6455f7d7f7-75pxj replicaset-controller logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:08 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6455f7d7f7 to 1 deployment-controller logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:09 +0000 UTC Normal Pod simple-prod-collector-77fcbdc546-mn6fm AddedInterface Add eth0 [10.129.2.76/23] from ovn-kubernetes logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:09 +0000 UTC Normal Pod simple-prod-collector-77fcbdc546-mn6fm.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" already present on machine kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:09 +0000 UTC Normal Pod simple-prod-collector-77fcbdc546-mn6fm.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:09 +0000 UTC Normal Pod simple-prod-collector-77fcbdc546-mn6fm.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:09 +0000 UTC Normal Pod simple-prod-query-6455f7d7f7-75pxj AddedInterface Add eth0 [10.129.2.77/23] from ovn-kubernetes logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:09 +0000 UTC Normal Pod simple-prod-query-6455f7d7f7-75pxj.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:09 +0000 UTC Normal Pod simple-prod-query-6455f7d7f7-75pxj.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:09 +0000 UTC Normal Pod simple-prod-query-6455f7d7f7-75pxj.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:09 +0000 UTC Normal Pod simple-prod-query-6455f7d7f7-75pxj.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:09 +0000 UTC Normal Pod simple-prod-query-6455f7d7f7-75pxj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:09 +0000 UTC Normal Pod simple-prod-query-6455f7d7f7-75pxj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:09 +0000 UTC Normal Pod simple-prod-query-6455f7d7f7-75pxj.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:09 +0000 UTC Normal Pod simple-prod-query-6455f7d7f7-75pxj.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:09 +0000 UTC Normal Pod simple-prod-query-6455f7d7f7-75pxj.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:14 +0000 UTC Normal Pod simple-prod-query-6455f7d7f7-75pxj.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:14 +0000 UTC Normal Pod simple-prod-query-6455f7d7f7-75pxj.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:14 +0000 UTC Normal Pod simple-prod-query-6455f7d7f7-75pxj.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:14 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6455f7d7f7 SuccessfulDelete Deleted pod: simple-prod-query-6455f7d7f7-75pxj replicaset-controller logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:14 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-6455f7d7f7 to 0 from 1 deployment-controller logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:15 +0000 UTC Normal Pod simple-prod-query-b785bc4f5-7c8cf Binding Scheduled Successfully assigned kuttl-test-top-wahoo/simple-prod-query-b785bc4f5-7c8cf to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:15 +0000 UTC Normal Pod simple-prod-query-b785bc4f5-7c8cf AddedInterface Add eth0 [10.129.2.78/23] from ovn-kubernetes logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:15 +0000 UTC Normal Pod simple-prod-query-b785bc4f5-7c8cf.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:15 +0000 UTC Normal Pod simple-prod-query-b785bc4f5-7c8cf.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:15 +0000 UTC Normal Pod simple-prod-query-b785bc4f5-7c8cf.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:15 +0000 UTC Normal Pod simple-prod-query-b785bc4f5-7c8cf.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:15 +0000 UTC Normal ReplicaSet.apps simple-prod-query-b785bc4f5 SuccessfulCreate Created pod: simple-prod-query-b785bc4f5-7c8cf replicaset-controller logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:15 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-b785bc4f5 to 1 deployment-controller logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:16 +0000 UTC Normal Pod simple-prod-query-b785bc4f5-7c8cf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:16 +0000 UTC Normal Pod simple-prod-query-b785bc4f5-7c8cf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:16 +0000 UTC Normal Pod simple-prod-query-b785bc4f5-7c8cf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:16 +0000 UTC Normal Pod simple-prod-query-b785bc4f5-7c8cf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:16 +0000 UTC Normal Pod simple-prod-query-b785bc4f5-7c8cf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:19 +0000 UTC Normal Pod check-span-qvdpl Binding Scheduled Successfully assigned kuttl-test-top-wahoo/check-span-qvdpl to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:19 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-qvdpl job-controller logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:19 +0000 UTC Normal Pod report-span-ff4rg Binding Scheduled Successfully assigned kuttl-test-top-wahoo/report-span-ff4rg to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:19 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-ff4rg job-controller logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:20 +0000 UTC Normal Pod check-span-qvdpl AddedInterface Add eth0 [10.131.0.53/23] from ovn-kubernetes logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:20 +0000 UTC Normal Pod check-span-qvdpl.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:20 +0000 UTC Normal Pod check-span-qvdpl.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:20 +0000 UTC Normal Pod check-span-qvdpl.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:20 +0000 UTC Normal Pod report-span-ff4rg AddedInterface Add eth0 [10.131.0.52/23] from ovn-kubernetes logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:20 +0000 UTC Normal Pod report-span-ff4rg.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:20 +0000 UTC Normal Pod report-span-ff4rg.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:20 +0000 UTC Normal Pod report-span-ff4rg.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:23 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:23 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:23 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:25:32 | examples-simple-prod | 2023-11-27 14:25:31 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:25:32 | examples-simple-prod | Deleting namespace: kuttl-test-top-wahoo === CONT kuttl/harness/examples-business-application-injected-sidecar logger.go:42: 14:25:43 | examples-business-application-injected-sidecar | Creating namespace: kuttl-test-sensible-asp logger.go:42: 14:25:43 | examples-business-application-injected-sidecar/0-install | starting test step 0-install logger.go:42: 14:25:43 | examples-business-application-injected-sidecar/0-install | Deployment:kuttl-test-sensible-asp/myapp created logger.go:42: 14:25:43 | examples-business-application-injected-sidecar/0-install | test step completed 0-install logger.go:42: 14:25:43 | examples-business-application-injected-sidecar/1-install | starting test step 1-install logger.go:42: 14:25:43 | examples-business-application-injected-sidecar/1-install | Jaeger:kuttl-test-sensible-asp/simplest created logger.go:42: 14:25:53 | examples-business-application-injected-sidecar/1-install | test step completed 1-install logger.go:42: 14:25:53 | examples-business-application-injected-sidecar/2-smoke-test | starting test step 2-smoke-test logger.go:42: 14:25:53 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 14:25:55 | examples-business-application-injected-sidecar/2-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:26:01 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 14:26:01 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:26:02 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/report-span created logger.go:42: 14:26:02 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/check-span created logger.go:42: 14:26:13 | examples-business-application-injected-sidecar/2-smoke-test | test step completed 2-smoke-test logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | examples-business-application-injected-sidecar events from ns kuttl-test-sensible-asp: logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:43 +0000 UTC Normal Pod myapp-679f79d5f8-27tsh Binding Scheduled Successfully assigned kuttl-test-sensible-asp/myapp-679f79d5f8-27tsh to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:43 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulCreate Created pod: myapp-679f79d5f8-27tsh replicaset-controller logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:43 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-679f79d5f8 to 1 deployment-controller logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:43 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-645d6867f7 to 1 deployment-controller logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:44 +0000 UTC Normal Pod myapp-645d6867f7-g6fbl Binding Scheduled Successfully assigned kuttl-test-sensible-asp/myapp-645d6867f7-g6fbl to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:44 +0000 UTC Warning Pod myapp-645d6867f7-g6fbl FailedMount MountVolume.SetUp failed for volume "simplest-trusted-ca" : configmap "simplest-trusted-ca" not found kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:44 +0000 UTC Warning Pod myapp-645d6867f7-g6fbl FailedMount MountVolume.SetUp failed for volume "simplest-service-ca" : configmap "simplest-service-ca" not found kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:44 +0000 UTC Normal ReplicaSet.apps myapp-645d6867f7 SuccessfulCreate Created pod: myapp-645d6867f7-g6fbl replicaset-controller logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:44 +0000 UTC Normal Pod myapp-679f79d5f8-27tsh AddedInterface Add eth0 [10.128.2.37/23] from ovn-kubernetes logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:44 +0000 UTC Normal Pod myapp-679f79d5f8-27tsh.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:47 +0000 UTC Normal Pod myapp-679f79d5f8-27tsh.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.647s (3.647s including waiting) kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:48 +0000 UTC Normal Pod myapp-679f79d5f8-27tsh.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:48 +0000 UTC Normal Pod myapp-679f79d5f8-27tsh.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:52 +0000 UTC Normal Pod simplest-6dd9f58df5-h69cd Binding Scheduled Successfully assigned kuttl-test-sensible-asp/simplest-6dd9f58df5-h69cd to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:52 +0000 UTC Normal Pod simplest-6dd9f58df5-h69cd AddedInterface Add eth0 [10.128.2.38/23] from ovn-kubernetes logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:52 +0000 UTC Normal Pod simplest-6dd9f58df5-h69cd.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:52 +0000 UTC Normal Pod simplest-6dd9f58df5-h69cd.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:52 +0000 UTC Normal Pod simplest-6dd9f58df5-h69cd.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:52 +0000 UTC Normal Pod simplest-6dd9f58df5-h69cd.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:52 +0000 UTC Normal Pod simplest-6dd9f58df5-h69cd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:52 +0000 UTC Normal Pod simplest-6dd9f58df5-h69cd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:52 +0000 UTC Normal ReplicaSet.apps simplest-6dd9f58df5 SuccessfulCreate Created pod: simplest-6dd9f58df5-h69cd replicaset-controller logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:52 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-6dd9f58df5 to 1 deployment-controller logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:55 +0000 UTC Warning Pod myapp-679f79d5f8-27tsh.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.128.2.37:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:59 +0000 UTC Normal Pod myapp-645d6867f7-g6fbl AddedInterface Add eth0 [10.129.2.79/23] from ovn-kubernetes logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:59 +0000 UTC Normal Pod simplest-6dd9f58df5-h69cd.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:59 +0000 UTC Normal Pod simplest-6dd9f58df5-h69cd.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:59 +0000 UTC Normal ReplicaSet.apps simplest-6dd9f58df5 SuccessfulDelete Deleted pod: simplest-6dd9f58df5-h69cd replicaset-controller logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:25:59 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-6dd9f58df5 to 0 from 1 deployment-controller logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:00 +0000 UTC Normal Pod myapp-645d6867f7-g6fbl.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:00 +0000 UTC Normal Pod simplest-7d58d9599b-67glh Binding Scheduled Successfully assigned kuttl-test-sensible-asp/simplest-7d58d9599b-67glh to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:00 +0000 UTC Normal Pod simplest-7d58d9599b-67glh AddedInterface Add eth0 [10.128.2.39/23] from ovn-kubernetes logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:00 +0000 UTC Normal Pod simplest-7d58d9599b-67glh.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:00 +0000 UTC Normal Pod simplest-7d58d9599b-67glh.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:00 +0000 UTC Normal Pod simplest-7d58d9599b-67glh.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:00 +0000 UTC Normal Pod simplest-7d58d9599b-67glh.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:00 +0000 UTC Normal Pod simplest-7d58d9599b-67glh.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:00 +0000 UTC Normal Pod simplest-7d58d9599b-67glh.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:00 +0000 UTC Normal ReplicaSet.apps simplest-7d58d9599b SuccessfulCreate Created pod: simplest-7d58d9599b-67glh replicaset-controller logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:00 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-7d58d9599b to 1 deployment-controller logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:02 +0000 UTC Normal Pod check-span-l5qfz Binding Scheduled Successfully assigned kuttl-test-sensible-asp/check-span-l5qfz to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:02 +0000 UTC Normal Pod check-span-l5qfz AddedInterface Add eth0 [10.131.0.55/23] from ovn-kubernetes logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:02 +0000 UTC Normal Pod check-span-l5qfz.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:02 +0000 UTC Normal Pod check-span-l5qfz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:02 +0000 UTC Normal Pod check-span-l5qfz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:02 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-l5qfz job-controller logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:02 +0000 UTC Normal Pod report-span-c66wm Binding Scheduled Successfully assigned kuttl-test-sensible-asp/report-span-c66wm to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:02 +0000 UTC Normal Pod report-span-c66wm AddedInterface Add eth0 [10.131.0.54/23] from ovn-kubernetes logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:02 +0000 UTC Normal Pod report-span-c66wm.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:02 +0000 UTC Normal Pod report-span-c66wm.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:02 +0000 UTC Normal Pod report-span-c66wm.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:02 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-c66wm job-controller logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:03 +0000 UTC Normal Pod myapp-645d6867f7-g6fbl.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.475s (3.475s including waiting) kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:03 +0000 UTC Normal Pod myapp-645d6867f7-g6fbl.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:03 +0000 UTC Normal Pod myapp-645d6867f7-g6fbl.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:03 +0000 UTC Normal Pod myapp-645d6867f7-g6fbl.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:03 +0000 UTC Normal Pod myapp-645d6867f7-g6fbl.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:03 +0000 UTC Normal Pod myapp-645d6867f7-g6fbl.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:04 +0000 UTC Normal Pod myapp-679f79d5f8-27tsh.spec.containers{myapp} Killing Stopping container myapp kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:04 +0000 UTC Warning Pod myapp-679f79d5f8-27tsh.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.128.2.37:8080/": EOF kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:04 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulDelete Deleted pod: myapp-679f79d5f8-27tsh replicaset-controller logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:04 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled down replica set myapp-679f79d5f8 to 0 from 1 deployment-controller logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:10 +0000 UTC Warning Pod myapp-645d6867f7-g6fbl.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.129.2.79:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | 2023-11-27 14:26:13 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:26:13 | examples-business-application-injected-sidecar | Deleting namespace: kuttl-test-sensible-asp === CONT kuttl/harness/examples-openshift-with-htpasswd logger.go:42: 14:26:24 | examples-openshift-with-htpasswd | Ignoring 00-install.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:26:24 | examples-openshift-with-htpasswd | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:26:24 | examples-openshift-with-htpasswd | Creating namespace: kuttl-test-touching-yak logger.go:42: 14:26:24 | examples-openshift-with-htpasswd/0-install | starting test step 0-install logger.go:42: 14:26:25 | examples-openshift-with-htpasswd/0-install | Secret:kuttl-test-touching-yak/htpasswd created logger.go:42: 14:26:25 | examples-openshift-with-htpasswd/0-install | test step completed 0-install logger.go:42: 14:26:25 | examples-openshift-with-htpasswd/1-install | starting test step 1-install logger.go:42: 14:26:25 | examples-openshift-with-htpasswd/1-install | Jaeger:kuttl-test-touching-yak/with-htpasswd created logger.go:42: 14:26:31 | examples-openshift-with-htpasswd/1-install | test step completed 1-install logger.go:42: 14:26:31 | examples-openshift-with-htpasswd/2-check-unsecured | starting test step 2-check-unsecured logger.go:42: 14:26:31 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [./ensure-ingress-host.sh] logger.go:42: 14:26:31 | examples-openshift-with-htpasswd/2-check-unsecured | Checking the Ingress host value was populated logger.go:42: 14:26:31 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 0 logger.go:42: 14:26:31 | examples-openshift-with-htpasswd/2-check-unsecured | Hostname is with-htpasswd-kuttl-test-touching-yak.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com logger.go:42: 14:26:31 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 14:26:31 | examples-openshift-with-htpasswd/2-check-unsecured | Checking an expected HTTP response logger.go:42: 14:26:31 | examples-openshift-with-htpasswd/2-check-unsecured | Running in OpenShift logger.go:42: 14:26:31 | examples-openshift-with-htpasswd/2-check-unsecured | Not using any secret logger.go:42: 14:26:31 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 1/30 the https://with-htpasswd-kuttl-test-touching-yak.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 14:26:31 | examples-openshift-with-htpasswd/2-check-unsecured | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 14:26:31 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 2/30 the https://with-htpasswd-kuttl-test-touching-yak.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 14:26:31 | examples-openshift-with-htpasswd/2-check-unsecured | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 3/30 the https://with-htpasswd-kuttl-test-touching-yak.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/2-check-unsecured | curl response asserted properly logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/2-check-unsecured | test step completed 2-check-unsecured logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/3-check-unauthorized | starting test step 3-check-unauthorized logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [./ensure-ingress-host.sh] logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking the Ingress host value was populated logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 0 logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/3-check-unauthorized | Hostname is with-htpasswd-kuttl-test-touching-yak.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [sh -c JAEGER_USERNAME=wronguser JAEGER_PASSWORD=wrongpassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking an expected HTTP response logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/3-check-unauthorized | Running in OpenShift logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/3-check-unauthorized | Using Jaeger basic authentication logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 1/30 the https://with-htpasswd-kuttl-test-touching-yak.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/3-check-unauthorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 2/30 the https://with-htpasswd-kuttl-test-touching-yak.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/3-check-unauthorized | curl response asserted properly logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/3-check-unauthorized | test step completed 3-check-unauthorized logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/4-check-authorized | starting test step 4-check-authorized logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/4-check-authorized | running command: [./ensure-ingress-host.sh] logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/4-check-authorized | Checking the Ingress host value was populated logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/4-check-authorized | Try number 0 logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/4-check-authorized | Hostname is with-htpasswd-kuttl-test-touching-yak.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/4-check-authorized | running command: [sh -c JAEGER_USERNAME=awesomeuser JAEGER_PASSWORD=awesomepassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE with-htpasswd] logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/4-check-authorized | Checking an expected HTTP response logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/4-check-authorized | Running in OpenShift logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/4-check-authorized | Using Jaeger basic authentication logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/4-check-authorized | Try number 1/30 the https://with-htpasswd-kuttl-test-touching-yak.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/4-check-authorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/4-check-authorized | Try number 2/30 the https://with-htpasswd-kuttl-test-touching-yak.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/4-check-authorized | curl response asserted properly logger.go:42: 14:26:41 | examples-openshift-with-htpasswd/4-check-authorized | test step completed 4-check-authorized logger.go:42: 14:26:41 | examples-openshift-with-htpasswd | examples-openshift-with-htpasswd events from ns kuttl-test-touching-yak: logger.go:42: 14:26:41 | examples-openshift-with-htpasswd | 2023-11-27 14:26:28 +0000 UTC Normal Pod with-htpasswd-5ff6898d6b-w6j5j Binding Scheduled Successfully assigned kuttl-test-touching-yak/with-htpasswd-5ff6898d6b-w6j5j to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:26:41 | examples-openshift-with-htpasswd | 2023-11-27 14:26:28 +0000 UTC Normal Pod with-htpasswd-5ff6898d6b-w6j5j AddedInterface Add eth0 [10.129.2.80/23] from ovn-kubernetes logger.go:42: 14:26:41 | examples-openshift-with-htpasswd | 2023-11-27 14:26:28 +0000 UTC Normal ReplicaSet.apps with-htpasswd-5ff6898d6b SuccessfulCreate Created pod: with-htpasswd-5ff6898d6b-w6j5j replicaset-controller logger.go:42: 14:26:41 | examples-openshift-with-htpasswd | 2023-11-27 14:26:28 +0000 UTC Normal Deployment.apps with-htpasswd ScalingReplicaSet Scaled up replica set with-htpasswd-5ff6898d6b to 1 deployment-controller logger.go:42: 14:26:41 | examples-openshift-with-htpasswd | 2023-11-27 14:26:29 +0000 UTC Normal Pod with-htpasswd-5ff6898d6b-w6j5j.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:26:41 | examples-openshift-with-htpasswd | 2023-11-27 14:26:29 +0000 UTC Normal Pod with-htpasswd-5ff6898d6b-w6j5j.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:26:41 | examples-openshift-with-htpasswd | 2023-11-27 14:26:29 +0000 UTC Normal Pod with-htpasswd-5ff6898d6b-w6j5j.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:26:41 | examples-openshift-with-htpasswd | 2023-11-27 14:26:29 +0000 UTC Normal Pod with-htpasswd-5ff6898d6b-w6j5j.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:26:41 | examples-openshift-with-htpasswd | 2023-11-27 14:26:29 +0000 UTC Normal Pod with-htpasswd-5ff6898d6b-w6j5j.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:26:41 | examples-openshift-with-htpasswd | 2023-11-27 14:26:29 +0000 UTC Normal Pod with-htpasswd-5ff6898d6b-w6j5j.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:26:41 | examples-openshift-with-htpasswd | Deleting namespace: kuttl-test-touching-yak === CONT kuttl/harness/examples-openshift-agent-as-daemonset logger.go:42: 14:26:47 | examples-openshift-agent-as-daemonset | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:26:47 | examples-openshift-agent-as-daemonset | Creating namespace: kuttl-test-fast-feline logger.go:42: 14:26:47 | examples-openshift-agent-as-daemonset/0-install | starting test step 0-install logger.go:42: 14:26:47 | examples-openshift-agent-as-daemonset/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 14:26:47 | examples-openshift-agent-as-daemonset/0-install | ServiceAccount:kuttl-test-fast-feline/jaeger-agent-daemonset created logger.go:42: 14:26:47 | examples-openshift-agent-as-daemonset/0-install | test step completed 0-install logger.go:42: 14:26:47 | examples-openshift-agent-as-daemonset/1-add-policy | starting test step 1-add-policy logger.go:42: 14:26:47 | examples-openshift-agent-as-daemonset/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 14:26:47 | examples-openshift-agent-as-daemonset/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 14:26:47 | examples-openshift-agent-as-daemonset/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 14:26:52 | examples-openshift-agent-as-daemonset/1-add-policy | test step completed 1-add-policy logger.go:42: 14:26:52 | examples-openshift-agent-as-daemonset/2-install | starting test step 2-install logger.go:42: 14:26:52 | examples-openshift-agent-as-daemonset/2-install | Jaeger:kuttl-test-fast-feline/agent-as-daemonset created logger.go:42: 14:26:59 | examples-openshift-agent-as-daemonset/2-install | test step completed 2-install logger.go:42: 14:26:59 | examples-openshift-agent-as-daemonset/3-install | starting test step 3-install logger.go:42: 14:26:59 | examples-openshift-agent-as-daemonset/3-install | Deployment:kuttl-test-fast-feline/vertx-create-span-sidecar created logger.go:42: 14:27:01 | examples-openshift-agent-as-daemonset/3-install | test step completed 3-install logger.go:42: 14:27:01 | examples-openshift-agent-as-daemonset/4-find-service | starting test step 4-find-service logger.go:42: 14:27:01 | examples-openshift-agent-as-daemonset/4-find-service | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 14:27:02 | examples-openshift-agent-as-daemonset/4-find-service | Warning: resource jaegers/agent-as-daemonset is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:27:08 | examples-openshift-agent-as-daemonset/4-find-service | running command: [sh -c SERVICE_NAME=order ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JOB_NUMBER=00 JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o find-service-00-job.yaml] logger.go:42: 14:27:09 | examples-openshift-agent-as-daemonset/4-find-service | running command: [sh -c kubectl create -f find-service-00-job.yaml -n $NAMESPACE] logger.go:42: 14:27:09 | examples-openshift-agent-as-daemonset/4-find-service | job.batch/00-find-service created logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset/4-find-service | test step completed 4-find-service logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | examples-openshift-agent-as-daemonset events from ns kuttl-test-fast-feline: logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:56 +0000 UTC Normal Pod agent-as-daemonset-7b857c6c7b-dc8qs Binding Scheduled Successfully assigned kuttl-test-fast-feline/agent-as-daemonset-7b857c6c7b-dc8qs to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:56 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-7b857c6c7b SuccessfulCreate Created pod: agent-as-daemonset-7b857c6c7b-dc8qs replicaset-controller logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:56 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-9s6v5 Binding Scheduled Successfully assigned kuttl-test-fast-feline/agent-as-daemonset-agent-daemonset-9s6v5 to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:56 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-t8x68 Binding Scheduled Successfully assigned kuttl-test-fast-feline/agent-as-daemonset-agent-daemonset-t8x68 to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:56 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-z72zl Binding Scheduled Successfully assigned kuttl-test-fast-feline/agent-as-daemonset-agent-daemonset-z72zl to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:56 +0000 UTC Normal DaemonSet.apps agent-as-daemonset-agent-daemonset SuccessfulCreate Created pod: agent-as-daemonset-agent-daemonset-t8x68 daemonset-controller logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:56 +0000 UTC Normal DaemonSet.apps agent-as-daemonset-agent-daemonset SuccessfulCreate Created pod: agent-as-daemonset-agent-daemonset-9s6v5 daemonset-controller logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:56 +0000 UTC Normal DaemonSet.apps agent-as-daemonset-agent-daemonset SuccessfulCreate Created pod: agent-as-daemonset-agent-daemonset-z72zl daemonset-controller logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:56 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-7b857c6c7b to 1 deployment-controller logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:57 +0000 UTC Normal Pod agent-as-daemonset-7b857c6c7b-dc8qs AddedInterface Add eth0 [10.129.2.81/23] from ovn-kubernetes logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:57 +0000 UTC Normal Pod agent-as-daemonset-7b857c6c7b-dc8qs.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:57 +0000 UTC Normal Pod agent-as-daemonset-7b857c6c7b-dc8qs.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:57 +0000 UTC Normal Pod agent-as-daemonset-7b857c6c7b-dc8qs.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:57 +0000 UTC Normal Pod agent-as-daemonset-7b857c6c7b-dc8qs.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:57 +0000 UTC Normal Pod agent-as-daemonset-7b857c6c7b-dc8qs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:57 +0000 UTC Normal Pod agent-as-daemonset-7b857c6c7b-dc8qs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:57 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-9s6v5 AddedInterface Add eth0 [10.128.2.40/23] from ovn-kubernetes logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:57 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-9s6v5.spec.containers{jaeger-agent-daemonset} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:57 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-t8x68 AddedInterface Add eth0 [10.131.0.56/23] from ovn-kubernetes logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:57 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-t8x68.spec.containers{jaeger-agent-daemonset} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:57 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-t8x68.spec.containers{jaeger-agent-daemonset} Created Created container jaeger-agent-daemonset kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:57 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-t8x68.spec.containers{jaeger-agent-daemonset} Started Started container jaeger-agent-daemonset kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:57 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-z72zl AddedInterface Add eth0 [10.129.2.82/23] from ovn-kubernetes logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:57 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-z72zl.spec.containers{jaeger-agent-daemonset} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:57 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-z72zl.spec.containers{jaeger-agent-daemonset} Created Created container jaeger-agent-daemonset kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:57 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-z72zl.spec.containers{jaeger-agent-daemonset} Started Started container jaeger-agent-daemonset kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:59 +0000 UTC Normal Pod vertx-create-span-sidecar-6c569f6fc6-pvr42 Binding Scheduled Successfully assigned kuttl-test-fast-feline/vertx-create-span-sidecar-6c569f6fc6-pvr42 to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:59 +0000 UTC Normal Pod vertx-create-span-sidecar-6c569f6fc6-pvr42 AddedInterface Add eth0 [10.128.2.41/23] from ovn-kubernetes logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:59 +0000 UTC Normal Pod vertx-create-span-sidecar-6c569f6fc6-pvr42.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:59 +0000 UTC Normal Pod vertx-create-span-sidecar-6c569f6fc6-pvr42.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:59 +0000 UTC Normal Pod vertx-create-span-sidecar-6c569f6fc6-pvr42.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:59 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6c569f6fc6 SuccessfulCreate Created pod: vertx-create-span-sidecar-6c569f6fc6-pvr42 replicaset-controller logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:26:59 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6c569f6fc6 to 1 deployment-controller logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:01 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-9s6v5.spec.containers{jaeger-agent-daemonset} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" in 4.371s (4.371s including waiting) kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:01 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-9s6v5.spec.containers{jaeger-agent-daemonset} Created Created container jaeger-agent-daemonset kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:01 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-9s6v5.spec.containers{jaeger-agent-daemonset} Started Started container jaeger-agent-daemonset kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:03 +0000 UTC Normal Pod agent-as-daemonset-7b857c6c7b-dc8qs.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:03 +0000 UTC Normal Pod agent-as-daemonset-7b857c6c7b-dc8qs.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:03 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-7b857c6c7b SuccessfulDelete Deleted pod: agent-as-daemonset-7b857c6c7b-dc8qs replicaset-controller logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:03 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled down replica set agent-as-daemonset-7b857c6c7b to 0 from 1 deployment-controller logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:04 +0000 UTC Normal Pod agent-as-daemonset-74f94b5f59-wcpmg Binding Scheduled Successfully assigned kuttl-test-fast-feline/agent-as-daemonset-74f94b5f59-wcpmg to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:04 +0000 UTC Normal Pod agent-as-daemonset-74f94b5f59-wcpmg AddedInterface Add eth0 [10.129.2.83/23] from ovn-kubernetes logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:04 +0000 UTC Normal Pod agent-as-daemonset-74f94b5f59-wcpmg.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:04 +0000 UTC Normal Pod agent-as-daemonset-74f94b5f59-wcpmg.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:04 +0000 UTC Normal Pod agent-as-daemonset-74f94b5f59-wcpmg.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:04 +0000 UTC Normal Pod agent-as-daemonset-74f94b5f59-wcpmg.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:04 +0000 UTC Normal Pod agent-as-daemonset-74f94b5f59-wcpmg.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:04 +0000 UTC Normal Pod agent-as-daemonset-74f94b5f59-wcpmg.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:04 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-74f94b5f59 SuccessfulCreate Created pod: agent-as-daemonset-74f94b5f59-wcpmg replicaset-controller logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:04 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-74f94b5f59 to 1 deployment-controller logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:07 +0000 UTC Warning Pod vertx-create-span-sidecar-6c569f6fc6-pvr42.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.41:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:07 +0000 UTC Warning Pod vertx-create-span-sidecar-6c569f6fc6-pvr42.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.41:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:09 +0000 UTC Normal Pod 00-find-service-xzbrt Binding Scheduled Successfully assigned kuttl-test-fast-feline/00-find-service-xzbrt to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:09 +0000 UTC Normal Pod 00-find-service-xzbrt AddedInterface Add eth0 [10.128.2.42/23] from ovn-kubernetes logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:09 +0000 UTC Normal Pod 00-find-service-xzbrt.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:09 +0000 UTC Normal Pod 00-find-service-xzbrt.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:09 +0000 UTC Normal Pod 00-find-service-xzbrt.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:09 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-xzbrt job-controller logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:09 +0000 UTC Normal Pod vertx-create-span-sidecar-6c569f6fc6-pvr42.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:10 +0000 UTC Warning Pod vertx-create-span-sidecar-6c569f6fc6-pvr42.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.41:8080/": read tcp 10.128.2.2:36150->10.128.2.41:8080: read: connection reset by peer kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:10 +0000 UTC Warning Pod vertx-create-span-sidecar-6c569f6fc6-pvr42.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.41:8080/": dial tcp 10.128.2.41:8080: connect: connection refused kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:21 +0000 UTC Warning Pod vertx-create-span-sidecar-6c569f6fc6-pvr42.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.41:8080/": read tcp 10.128.2.2:34892->10.128.2.41:8080: read: connection reset by peer kubelet logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | 2023-11-27 14:27:44 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 14:27:44 | examples-openshift-agent-as-daemonset | Deleting namespace: kuttl-test-fast-feline === CONT kuttl/harness/examples-collector-with-priority-class logger.go:42: 14:27:50 | examples-collector-with-priority-class | Creating namespace: kuttl-test-good-alpaca logger.go:42: 14:27:50 | examples-collector-with-priority-class/0-install | starting test step 0-install logger.go:42: 14:27:50 | examples-collector-with-priority-class/0-install | PriorityClass:/collector-high-priority created logger.go:42: 14:27:50 | examples-collector-with-priority-class/0-install | Jaeger:kuttl-test-good-alpaca/collector-with-high-priority created logger.go:42: 14:27:56 | examples-collector-with-priority-class/0-install | test step completed 0-install logger.go:42: 14:27:56 | examples-collector-with-priority-class/1-smoke-test | starting test step 1-smoke-test logger.go:42: 14:27:56 | examples-collector-with-priority-class/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE collector-with-high-priority /dev/null] logger.go:42: 14:27:58 | examples-collector-with-priority-class/1-smoke-test | Warning: resource jaegers/collector-with-high-priority is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:28:04 | examples-collector-with-priority-class/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 14:28:05 | examples-collector-with-priority-class/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:28:05 | examples-collector-with-priority-class/1-smoke-test | job.batch/report-span created logger.go:42: 14:28:05 | examples-collector-with-priority-class/1-smoke-test | job.batch/check-span created logger.go:42: 14:28:16 | examples-collector-with-priority-class/1-smoke-test | test step completed 1-smoke-test logger.go:42: 14:28:16 | examples-collector-with-priority-class | examples-collector-with-priority-class events from ns kuttl-test-good-alpaca: logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:27:54 +0000 UTC Normal Pod collector-with-high-priority-6457d7844f-z55pb Binding Scheduled Successfully assigned kuttl-test-good-alpaca/collector-with-high-priority-6457d7844f-z55pb to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:27:54 +0000 UTC Normal Pod collector-with-high-priority-6457d7844f-z55pb AddedInterface Add eth0 [10.129.2.84/23] from ovn-kubernetes logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:27:54 +0000 UTC Normal Pod collector-with-high-priority-6457d7844f-z55pb.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:27:54 +0000 UTC Normal Pod collector-with-high-priority-6457d7844f-z55pb.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:27:54 +0000 UTC Normal Pod collector-with-high-priority-6457d7844f-z55pb.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:27:54 +0000 UTC Normal Pod collector-with-high-priority-6457d7844f-z55pb.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:27:54 +0000 UTC Normal Pod collector-with-high-priority-6457d7844f-z55pb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:27:54 +0000 UTC Normal Pod collector-with-high-priority-6457d7844f-z55pb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:27:54 +0000 UTC Normal ReplicaSet.apps collector-with-high-priority-6457d7844f SuccessfulCreate Created pod: collector-with-high-priority-6457d7844f-z55pb replicaset-controller logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:27:54 +0000 UTC Normal Deployment.apps collector-with-high-priority ScalingReplicaSet Scaled up replica set collector-with-high-priority-6457d7844f to 1 deployment-controller logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:02 +0000 UTC Normal Pod collector-with-high-priority-6457d7844f-z55pb.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:02 +0000 UTC Normal Pod collector-with-high-priority-6457d7844f-z55pb.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:02 +0000 UTC Normal ReplicaSet.apps collector-with-high-priority-6457d7844f SuccessfulDelete Deleted pod: collector-with-high-priority-6457d7844f-z55pb replicaset-controller logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:02 +0000 UTC Normal Pod collector-with-high-priority-7b4bb95fb4-xtkj2 Binding Scheduled Successfully assigned kuttl-test-good-alpaca/collector-with-high-priority-7b4bb95fb4-xtkj2 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:02 +0000 UTC Normal ReplicaSet.apps collector-with-high-priority-7b4bb95fb4 SuccessfulCreate Created pod: collector-with-high-priority-7b4bb95fb4-xtkj2 replicaset-controller logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:02 +0000 UTC Normal Deployment.apps collector-with-high-priority ScalingReplicaSet Scaled down replica set collector-with-high-priority-6457d7844f to 0 from 1 deployment-controller logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:02 +0000 UTC Normal Deployment.apps collector-with-high-priority ScalingReplicaSet Scaled up replica set collector-with-high-priority-7b4bb95fb4 to 1 deployment-controller logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:03 +0000 UTC Normal Pod collector-with-high-priority-7b4bb95fb4-xtkj2 AddedInterface Add eth0 [10.129.2.85/23] from ovn-kubernetes logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:03 +0000 UTC Normal Pod collector-with-high-priority-7b4bb95fb4-xtkj2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:03 +0000 UTC Normal Pod collector-with-high-priority-7b4bb95fb4-xtkj2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:03 +0000 UTC Normal Pod collector-with-high-priority-7b4bb95fb4-xtkj2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:03 +0000 UTC Normal Pod collector-with-high-priority-7b4bb95fb4-xtkj2.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:03 +0000 UTC Normal Pod collector-with-high-priority-7b4bb95fb4-xtkj2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:03 +0000 UTC Normal Pod collector-with-high-priority-7b4bb95fb4-xtkj2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:05 +0000 UTC Normal Pod check-span-m2jxm Binding Scheduled Successfully assigned kuttl-test-good-alpaca/check-span-m2jxm to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:05 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-m2jxm job-controller logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:05 +0000 UTC Normal Pod report-span-jlmtc Binding Scheduled Successfully assigned kuttl-test-good-alpaca/report-span-jlmtc to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:05 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-jlmtc job-controller logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:06 +0000 UTC Normal Pod check-span-m2jxm AddedInterface Add eth0 [10.131.0.57/23] from ovn-kubernetes logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:06 +0000 UTC Normal Pod check-span-m2jxm.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:06 +0000 UTC Normal Pod check-span-m2jxm.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:06 +0000 UTC Normal Pod check-span-m2jxm.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:06 +0000 UTC Normal Pod report-span-jlmtc AddedInterface Add eth0 [10.128.2.43/23] from ovn-kubernetes logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:06 +0000 UTC Normal Pod report-span-jlmtc.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:06 +0000 UTC Normal Pod report-span-jlmtc.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:06 +0000 UTC Normal Pod report-span-jlmtc.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:28:16 | examples-collector-with-priority-class | 2023-11-27 14:28:16 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:28:16 | examples-collector-with-priority-class | Deleting namespace: kuttl-test-good-alpaca === CONT kuttl/harness/examples-all-in-one-with-options logger.go:42: 14:28:23 | examples-all-in-one-with-options | Creating namespace: kuttl-test-genuine-panther logger.go:42: 14:28:23 | examples-all-in-one-with-options/0-install | starting test step 0-install logger.go:42: 14:28:23 | examples-all-in-one-with-options/0-install | Jaeger:kuttl-test-genuine-panther/my-jaeger created logger.go:42: 14:28:29 | examples-all-in-one-with-options/0-install | test step completed 0-install logger.go:42: 14:28:29 | examples-all-in-one-with-options/1-smoke-test | starting test step 1-smoke-test logger.go:42: 14:28:29 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 14:28:30 | examples-all-in-one-with-options/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:28:37 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443/jaeger MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 14:28:37 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:28:37 | examples-all-in-one-with-options/1-smoke-test | job.batch/report-span created logger.go:42: 14:28:38 | examples-all-in-one-with-options/1-smoke-test | job.batch/check-span created logger.go:42: 14:28:50 | examples-all-in-one-with-options/1-smoke-test | test step completed 1-smoke-test logger.go:42: 14:28:50 | examples-all-in-one-with-options | examples-all-in-one-with-options events from ns kuttl-test-genuine-panther: logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:27 +0000 UTC Normal Pod my-jaeger-58cbb44488-q274r Binding Scheduled Successfully assigned kuttl-test-genuine-panther/my-jaeger-58cbb44488-q274r to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:27 +0000 UTC Warning Pod my-jaeger-58cbb44488-q274r FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:27 +0000 UTC Normal ReplicaSet.apps my-jaeger-58cbb44488 SuccessfulCreate Created pod: my-jaeger-58cbb44488-q274r replicaset-controller logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:27 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-58cbb44488 to 1 deployment-controller logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:28 +0000 UTC Normal Pod my-jaeger-58cbb44488-q274r AddedInterface Add eth0 [10.128.2.44/23] from ovn-kubernetes logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:28 +0000 UTC Normal Pod my-jaeger-58cbb44488-q274r.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:28 +0000 UTC Normal Pod my-jaeger-58cbb44488-q274r.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:28 +0000 UTC Normal Pod my-jaeger-58cbb44488-q274r.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:28 +0000 UTC Normal Pod my-jaeger-58cbb44488-q274r.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:28 +0000 UTC Normal Pod my-jaeger-58cbb44488-q274r.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:28 +0000 UTC Normal Pod my-jaeger-58cbb44488-q274r.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:33 +0000 UTC Normal Pod my-jaeger-58cbb44488-q274r.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:33 +0000 UTC Normal Pod my-jaeger-58cbb44488-q274r.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:33 +0000 UTC Normal ReplicaSet.apps my-jaeger-58cbb44488 SuccessfulDelete Deleted pod: my-jaeger-58cbb44488-q274r replicaset-controller logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:33 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-58cbb44488 to 0 from 1 deployment-controller logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:34 +0000 UTC Normal Pod my-jaeger-5c9cc87d47-g8qdk Binding Scheduled Successfully assigned kuttl-test-genuine-panther/my-jaeger-5c9cc87d47-g8qdk to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:34 +0000 UTC Normal Pod my-jaeger-5c9cc87d47-g8qdk AddedInterface Add eth0 [10.128.2.45/23] from ovn-kubernetes logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:34 +0000 UTC Normal Pod my-jaeger-5c9cc87d47-g8qdk.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:34 +0000 UTC Normal ReplicaSet.apps my-jaeger-5c9cc87d47 SuccessfulCreate Created pod: my-jaeger-5c9cc87d47-g8qdk replicaset-controller logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:34 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5c9cc87d47 to 1 deployment-controller logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:35 +0000 UTC Normal Pod my-jaeger-5c9cc87d47-g8qdk.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:35 +0000 UTC Normal Pod my-jaeger-5c9cc87d47-g8qdk.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:35 +0000 UTC Normal Pod my-jaeger-5c9cc87d47-g8qdk.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:35 +0000 UTC Normal Pod my-jaeger-5c9cc87d47-g8qdk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:35 +0000 UTC Normal Pod my-jaeger-5c9cc87d47-g8qdk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:37 +0000 UTC Normal Pod report-span-b8fhp Binding Scheduled Successfully assigned kuttl-test-genuine-panther/report-span-b8fhp to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:37 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-b8fhp job-controller logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:38 +0000 UTC Normal Pod check-span-knv28 Binding Scheduled Successfully assigned kuttl-test-genuine-panther/check-span-knv28 to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:38 +0000 UTC Normal Pod check-span-knv28 AddedInterface Add eth0 [10.131.0.58/23] from ovn-kubernetes logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:38 +0000 UTC Normal Pod check-span-knv28.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:38 +0000 UTC Normal Pod check-span-knv28.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:38 +0000 UTC Normal Pod check-span-knv28.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:38 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-knv28 job-controller logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:38 +0000 UTC Normal Pod report-span-b8fhp AddedInterface Add eth0 [10.129.2.86/23] from ovn-kubernetes logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:38 +0000 UTC Normal Pod report-span-b8fhp.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:38 +0000 UTC Normal Pod report-span-b8fhp.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:38 +0000 UTC Normal Pod report-span-b8fhp.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:28:50 | examples-all-in-one-with-options | 2023-11-27 14:28:49 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:28:50 | examples-all-in-one-with-options | Deleting namespace: kuttl-test-genuine-panther === CONT kuttl/harness/examples-auto-provision-kafka logger.go:42: 14:29:01 | examples-auto-provision-kafka | Creating namespace: kuttl-test-exotic-flamingo logger.go:42: 14:29:01 | examples-auto-provision-kafka/2-install | starting test step 2-install logger.go:42: 14:29:01 | examples-auto-provision-kafka/2-install | Jaeger:kuttl-test-exotic-flamingo/auto-provision-kafka created logger.go:42: 14:30:10 | examples-auto-provision-kafka/2-install | test step completed 2-install logger.go:42: 14:30:10 | examples-auto-provision-kafka/3- | starting test step 3- logger.go:42: 14:30:41 | examples-auto-provision-kafka/3- | test step completed 3- logger.go:42: 14:30:41 | examples-auto-provision-kafka/4- | starting test step 4- logger.go:42: 14:31:02 | examples-auto-provision-kafka/4- | test step completed 4- logger.go:42: 14:31:02 | examples-auto-provision-kafka/5- | starting test step 5- logger.go:42: 14:31:07 | examples-auto-provision-kafka/5- | test step completed 5- logger.go:42: 14:31:07 | examples-auto-provision-kafka/6-smoke-test | starting test step 6-smoke-test logger.go:42: 14:31:07 | examples-auto-provision-kafka/6-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE auto-provision-kafka /dev/null] logger.go:42: 14:31:09 | examples-auto-provision-kafka/6-smoke-test | Warning: resource jaegers/auto-provision-kafka is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:31:15 | examples-auto-provision-kafka/6-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JAEGER_COLLECTOR_ENDPOINT=http://auto-provision-kafka-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://auto-provision-kafka-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 14:31:15 | examples-auto-provision-kafka/6-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:31:16 | examples-auto-provision-kafka/6-smoke-test | job.batch/report-span created logger.go:42: 14:31:16 | examples-auto-provision-kafka/6-smoke-test | job.batch/check-span created logger.go:42: 14:31:27 | examples-auto-provision-kafka/6-smoke-test | test step completed 6-smoke-test logger.go:42: 14:31:27 | examples-auto-provision-kafka | examples-auto-provision-kafka events from ns kuttl-test-exotic-flamingo: logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:07 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestexoticflamingoautoprovisionk-1-6xqc2 Binding Scheduled Successfully assigned kuttl-test-exotic-flamingo/elasticsearch-cdm-kuttltestexoticflamingoautoprovisionk-1-6xqc2 to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:07 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestexoticflamingoautoprovisionk-1-788f44b84d SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestexoticflamingoautoprovisionk-1-6xqc2 replicaset-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:07 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestexoticflamingoautoprovisionk-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestexoticflamingoautoprovisionk-1-788f44b84d to 1 deployment-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestexoticflamingoautoprovisionk-1-6xqc2 AddedInterface Add eth0 [10.128.2.46/23] from ovn-kubernetes logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestexoticflamingoautoprovisionk-1-6xqc2.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:56ea62bfb0ca36e19a7b21aff3676e49511f05f72da5e76d6427fd8240f328a8" already present on machine kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestexoticflamingoautoprovisionk-1-6xqc2.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestexoticflamingoautoprovisionk-1-6xqc2.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestexoticflamingoautoprovisionk-1-6xqc2.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:79427bea6b5c37894b9782c3821d8b9074838e606daa4a743b2ae060856fa98a" already present on machine kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestexoticflamingoautoprovisionk-1-6xqc2.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestexoticflamingoautoprovisionk-1-6xqc2.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:23 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestexoticflamingoautoprovisionk-1-6xqc2.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:34 +0000 UTC Normal PodDisruptionBudget.policy auto-provision-kafka-zookeeper NoPods No matching pods found controllermanager logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:34 +0000 UTC Normal PersistentVolumeClaim data-auto-provision-kafka-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:35 +0000 UTC Normal PersistentVolumeClaim data-auto-provision-kafka-zookeeper-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-exotic-flamingo/data-auto-provision-kafka-zookeeper-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-684659d4f7-2wwfg_180f48e7-ecaa-4b6b-8bcf-6bd94c7ad5a7 logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:35 +0000 UTC Normal PersistentVolumeClaim data-auto-provision-kafka-zookeeper-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:38 +0000 UTC Normal PersistentVolumeClaim data-auto-provision-kafka-zookeeper-0 ProvisioningSucceeded Successfully provisioned volume pvc-e4889b8d-af4d-4740-989c-e11ab94b223d ebs.csi.aws.com_aws-ebs-csi-driver-controller-684659d4f7-2wwfg_180f48e7-ecaa-4b6b-8bcf-6bd94c7ad5a7 logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:39 +0000 UTC Normal Pod auto-provision-kafka-zookeeper-0 Binding Scheduled Successfully assigned kuttl-test-exotic-flamingo/auto-provision-kafka-zookeeper-0 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:41 +0000 UTC Normal Pod auto-provision-kafka-zookeeper-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-e4889b8d-af4d-4740-989c-e11ab94b223d" attachdetach-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:49 +0000 UTC Normal Pod auto-provision-kafka-zookeeper-0 AddedInterface Add eth0 [10.129.2.87/23] from ovn-kubernetes logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:49 +0000 UTC Normal Pod auto-provision-kafka-zookeeper-0.spec.containers{zookeeper} Pulled Container image "registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:0bcdd55f01638f650ed69ebdf5f8a7291e103805b8cbb34013ced88e46e0678c" already present on machine kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:50 +0000 UTC Normal Pod auto-provision-kafka-zookeeper-0.spec.containers{zookeeper} Created Created container zookeeper kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:29:50 +0000 UTC Normal Pod auto-provision-kafka-zookeeper-0.spec.containers{zookeeper} Started Started container zookeeper kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:11 +0000 UTC Normal PodDisruptionBudget.policy auto-provision-kafka-kafka NoPods No matching pods found controllermanager logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:11 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provision-kafka-kafka-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:11 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provision-kafka-kafka-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:11 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provision-kafka-kafka-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-exotic-flamingo/data-0-auto-provision-kafka-kafka-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-684659d4f7-2wwfg_180f48e7-ecaa-4b6b-8bcf-6bd94c7ad5a7 logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:14 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provision-kafka-kafka-0 ProvisioningSucceeded Successfully provisioned volume pvc-0afcb101-e74e-498e-823b-b486c48425c4 ebs.csi.aws.com_aws-ebs-csi-driver-controller-684659d4f7-2wwfg_180f48e7-ecaa-4b6b-8bcf-6bd94c7ad5a7 logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:15 +0000 UTC Normal Pod auto-provision-kafka-kafka-0 Binding Scheduled Successfully assigned kuttl-test-exotic-flamingo/auto-provision-kafka-kafka-0 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:17 +0000 UTC Normal Pod auto-provision-kafka-kafka-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-0afcb101-e74e-498e-823b-b486c48425c4" attachdetach-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:20 +0000 UTC Normal Pod auto-provision-kafka-kafka-0 AddedInterface Add eth0 [10.129.2.89/23] from ovn-kubernetes logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:20 +0000 UTC Normal Pod auto-provision-kafka-kafka-0.spec.containers{kafka} Pulled Container image "registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:0bcdd55f01638f650ed69ebdf5f8a7291e103805b8cbb34013ced88e46e0678c" already present on machine kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:20 +0000 UTC Normal Pod auto-provision-kafka-kafka-0.spec.containers{kafka} Created Created container kafka kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:20 +0000 UTC Normal Pod auto-provision-kafka-kafka-0.spec.containers{kafka} Started Started container kafka kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:41 +0000 UTC Normal Pod auto-provision-kafka-entity-operator-75567b4589-496ll Binding Scheduled Successfully assigned kuttl-test-exotic-flamingo/auto-provision-kafka-entity-operator-75567b4589-496ll to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:41 +0000 UTC Normal Pod auto-provision-kafka-entity-operator-75567b4589-496ll AddedInterface Add eth0 [10.129.2.90/23] from ovn-kubernetes logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:41 +0000 UTC Normal Pod auto-provision-kafka-entity-operator-75567b4589-496ll.spec.containers{topic-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:464b04e622e0b3472e8a1e1ce8a2efd32cf27fc2056d3d589bfe6b5f9ac0bf4e" already present on machine kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:41 +0000 UTC Normal Pod auto-provision-kafka-entity-operator-75567b4589-496ll.spec.containers{topic-operator} Created Created container topic-operator kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:41 +0000 UTC Normal Pod auto-provision-kafka-entity-operator-75567b4589-496ll.spec.containers{topic-operator} Started Started container topic-operator kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:41 +0000 UTC Normal Pod auto-provision-kafka-entity-operator-75567b4589-496ll.spec.containers{user-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:464b04e622e0b3472e8a1e1ce8a2efd32cf27fc2056d3d589bfe6b5f9ac0bf4e" already present on machine kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:41 +0000 UTC Normal Pod auto-provision-kafka-entity-operator-75567b4589-496ll.spec.containers{user-operator} Created Created container user-operator kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:41 +0000 UTC Normal Pod auto-provision-kafka-entity-operator-75567b4589-496ll.spec.containers{user-operator} Started Started container user-operator kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:41 +0000 UTC Normal Pod auto-provision-kafka-entity-operator-75567b4589-496ll.spec.containers{tls-sidecar} Pulled Container image "registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:0bcdd55f01638f650ed69ebdf5f8a7291e103805b8cbb34013ced88e46e0678c" already present on machine kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:41 +0000 UTC Normal ReplicaSet.apps auto-provision-kafka-entity-operator-75567b4589 SuccessfulCreate Created pod: auto-provision-kafka-entity-operator-75567b4589-496ll replicaset-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:41 +0000 UTC Normal Deployment.apps auto-provision-kafka-entity-operator ScalingReplicaSet Scaled up replica set auto-provision-kafka-entity-operator-75567b4589 to 1 deployment-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:42 +0000 UTC Normal Pod auto-provision-kafka-entity-operator-75567b4589-496ll.spec.containers{tls-sidecar} Created Created container tls-sidecar kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:30:42 +0000 UTC Normal Pod auto-provision-kafka-entity-operator-75567b4589-496ll.spec.containers{tls-sidecar} Started Started container tls-sidecar kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:04 +0000 UTC Normal Pod auto-provision-kafka-collector-7dc8765d6b-5g6df Binding Scheduled Successfully assigned kuttl-test-exotic-flamingo/auto-provision-kafka-collector-7dc8765d6b-5g6df to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:04 +0000 UTC Normal Pod auto-provision-kafka-collector-7dc8765d6b-5g6df AddedInterface Add eth0 [10.131.0.59/23] from ovn-kubernetes logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:04 +0000 UTC Normal Pod auto-provision-kafka-collector-7dc8765d6b-5g6df.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" already present on machine kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:04 +0000 UTC Normal Pod auto-provision-kafka-collector-7dc8765d6b-5g6df.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:04 +0000 UTC Normal ReplicaSet.apps auto-provision-kafka-collector-7dc8765d6b SuccessfulCreate Created pod: auto-provision-kafka-collector-7dc8765d6b-5g6df replicaset-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:04 +0000 UTC Normal Deployment.apps auto-provision-kafka-collector ScalingReplicaSet Scaled up replica set auto-provision-kafka-collector-7dc8765d6b to 1 deployment-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:04 +0000 UTC Normal Pod auto-provision-kafka-ingester-599fbbb77d-v55td Binding Scheduled Successfully assigned kuttl-test-exotic-flamingo/auto-provision-kafka-ingester-599fbbb77d-v55td to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:04 +0000 UTC Normal Pod auto-provision-kafka-ingester-599fbbb77d-v55td AddedInterface Add eth0 [10.131.0.60/23] from ovn-kubernetes logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:04 +0000 UTC Normal Pod auto-provision-kafka-ingester-599fbbb77d-v55td.spec.containers{jaeger-ingester} Pulled Container image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:109a3676bfaccda0021c57bbe82bceed5140faaedb8bad4d7d2be7cd660de039" already present on machine kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:04 +0000 UTC Normal ReplicaSet.apps auto-provision-kafka-ingester-599fbbb77d SuccessfulCreate Created pod: auto-provision-kafka-ingester-599fbbb77d-v55td replicaset-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:04 +0000 UTC Normal Deployment.apps auto-provision-kafka-ingester ScalingReplicaSet Scaled up replica set auto-provision-kafka-ingester-599fbbb77d to 1 deployment-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:04 +0000 UTC Normal Pod auto-provision-kafka-query-696bdf9887-4wwp4 Binding Scheduled Successfully assigned kuttl-test-exotic-flamingo/auto-provision-kafka-query-696bdf9887-4wwp4 to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:04 +0000 UTC Normal Pod auto-provision-kafka-query-696bdf9887-4wwp4 AddedInterface Add eth0 [10.131.0.61/23] from ovn-kubernetes logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:04 +0000 UTC Normal Pod auto-provision-kafka-query-696bdf9887-4wwp4.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:04 +0000 UTC Normal ReplicaSet.apps auto-provision-kafka-query-696bdf9887 SuccessfulCreate Created pod: auto-provision-kafka-query-696bdf9887-4wwp4 replicaset-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:04 +0000 UTC Normal Deployment.apps auto-provision-kafka-query ScalingReplicaSet Scaled up replica set auto-provision-kafka-query-696bdf9887 to 1 deployment-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:05 +0000 UTC Normal Pod auto-provision-kafka-collector-7dc8765d6b-5g6df.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:05 +0000 UTC Normal Pod auto-provision-kafka-ingester-599fbbb77d-v55td.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:05 +0000 UTC Normal Pod auto-provision-kafka-ingester-599fbbb77d-v55td.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:05 +0000 UTC Normal Pod auto-provision-kafka-query-696bdf9887-4wwp4.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:05 +0000 UTC Normal Pod auto-provision-kafka-query-696bdf9887-4wwp4.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:05 +0000 UTC Normal Pod auto-provision-kafka-query-696bdf9887-4wwp4.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:05 +0000 UTC Normal Pod auto-provision-kafka-query-696bdf9887-4wwp4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:05 +0000 UTC Normal Pod auto-provision-kafka-query-696bdf9887-4wwp4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:05 +0000 UTC Normal Pod auto-provision-kafka-query-696bdf9887-4wwp4.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:05 +0000 UTC Normal Pod auto-provision-kafka-query-696bdf9887-4wwp4.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:05 +0000 UTC Normal Pod auto-provision-kafka-query-696bdf9887-4wwp4.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:11 +0000 UTC Normal Pod auto-provision-kafka-query-696bdf9887-4wwp4.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:11 +0000 UTC Normal Pod auto-provision-kafka-query-696bdf9887-4wwp4.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:11 +0000 UTC Normal Pod auto-provision-kafka-query-696bdf9887-4wwp4.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:11 +0000 UTC Normal ReplicaSet.apps auto-provision-kafka-query-696bdf9887 SuccessfulDelete Deleted pod: auto-provision-kafka-query-696bdf9887-4wwp4 replicaset-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:11 +0000 UTC Normal Deployment.apps auto-provision-kafka-query ScalingReplicaSet Scaled down replica set auto-provision-kafka-query-696bdf9887 to 0 from 1 deployment-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:12 +0000 UTC Normal Pod auto-provision-kafka-query-5cf6d9dc5-nl8tc Binding Scheduled Successfully assigned kuttl-test-exotic-flamingo/auto-provision-kafka-query-5cf6d9dc5-nl8tc to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:12 +0000 UTC Normal Pod auto-provision-kafka-query-5cf6d9dc5-nl8tc AddedInterface Add eth0 [10.131.0.62/23] from ovn-kubernetes logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:12 +0000 UTC Normal Pod auto-provision-kafka-query-5cf6d9dc5-nl8tc.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:12 +0000 UTC Normal Pod auto-provision-kafka-query-5cf6d9dc5-nl8tc.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:12 +0000 UTC Normal Pod auto-provision-kafka-query-5cf6d9dc5-nl8tc.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:12 +0000 UTC Normal Pod auto-provision-kafka-query-5cf6d9dc5-nl8tc.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:12 +0000 UTC Normal Pod auto-provision-kafka-query-5cf6d9dc5-nl8tc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:12 +0000 UTC Normal Pod auto-provision-kafka-query-5cf6d9dc5-nl8tc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:12 +0000 UTC Normal Pod auto-provision-kafka-query-5cf6d9dc5-nl8tc.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:12 +0000 UTC Normal Pod auto-provision-kafka-query-5cf6d9dc5-nl8tc.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:12 +0000 UTC Normal Pod auto-provision-kafka-query-5cf6d9dc5-nl8tc.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:12 +0000 UTC Normal ReplicaSet.apps auto-provision-kafka-query-5cf6d9dc5 SuccessfulCreate Created pod: auto-provision-kafka-query-5cf6d9dc5-nl8tc replicaset-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:12 +0000 UTC Normal Deployment.apps auto-provision-kafka-query ScalingReplicaSet Scaled up replica set auto-provision-kafka-query-5cf6d9dc5 to 1 deployment-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:16 +0000 UTC Normal Pod check-span-bcb87 Binding Scheduled Successfully assigned kuttl-test-exotic-flamingo/check-span-bcb87 to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:16 +0000 UTC Normal Pod check-span-bcb87 AddedInterface Add eth0 [10.131.0.63/23] from ovn-kubernetes logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:16 +0000 UTC Normal Pod check-span-bcb87.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:16 +0000 UTC Normal Pod check-span-bcb87.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:16 +0000 UTC Normal Pod check-span-bcb87.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:16 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-bcb87 job-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:16 +0000 UTC Normal Pod report-span-jfqsq Binding Scheduled Successfully assigned kuttl-test-exotic-flamingo/report-span-jfqsq to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:16 +0000 UTC Normal Pod report-span-jfqsq AddedInterface Add eth0 [10.129.2.91/23] from ovn-kubernetes logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:16 +0000 UTC Normal Pod report-span-jfqsq.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:16 +0000 UTC Normal Pod report-span-jfqsq.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:16 +0000 UTC Normal Pod report-span-jfqsq.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:16 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-jfqsq job-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provision-kafka-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provision-kafka-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provision-kafka-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provision-kafka-ingester FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provision-kafka-ingester FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provision-kafka-ingester FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:31:27 | examples-auto-provision-kafka | 2023-11-27 14:31:26 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:31:27 | examples-auto-provision-kafka | Deleting namespace: kuttl-test-exotic-flamingo === CONT kuttl/harness/examples-agent-with-priority-class logger.go:42: 14:32:07 | examples-agent-with-priority-class | Creating namespace: kuttl-test-amusing-sunbird logger.go:42: 14:32:07 | examples-agent-with-priority-class/0-install | starting test step 0-install logger.go:42: 14:32:07 | examples-agent-with-priority-class/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 14:32:07 | examples-agent-with-priority-class/0-install | ServiceAccount:kuttl-test-amusing-sunbird/jaeger-agent-daemonset created logger.go:42: 14:32:07 | examples-agent-with-priority-class/0-install | test step completed 0-install logger.go:42: 14:32:07 | examples-agent-with-priority-class/1-install | starting test step 1-install logger.go:42: 14:32:07 | examples-agent-with-priority-class/1-install | PriorityClass:/high-priority created logger.go:42: 14:32:07 | examples-agent-with-priority-class/1-install | Jaeger:kuttl-test-amusing-sunbird/agent-as-daemonset created logger.go:42: 14:32:13 | examples-agent-with-priority-class/1-install | test step completed 1-install logger.go:42: 14:32:13 | examples-agent-with-priority-class/2-smoke-test | starting test step 2-smoke-test logger.go:42: 14:32:13 | examples-agent-with-priority-class/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 14:32:15 | examples-agent-with-priority-class/2-smoke-test | Warning: resource jaegers/agent-as-daemonset is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:32:21 | examples-agent-with-priority-class/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 14:32:21 | examples-agent-with-priority-class/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:32:22 | examples-agent-with-priority-class/2-smoke-test | job.batch/report-span created logger.go:42: 14:32:22 | examples-agent-with-priority-class/2-smoke-test | job.batch/check-span created logger.go:42: 14:42:22 | examples-agent-with-priority-class/2-smoke-test | test step failed 2-smoke-test case.go:364: failed in step 2-smoke-test case.go:366: --- Job:kuttl-test-amusing-sunbird/check-span +++ Job:kuttl-test-amusing-sunbird/check-span @@ -1,8 +1,141 @@ apiVersion: batch/v1 kind: Job metadata: + annotations: + kubectl.kubernetes.io/last-applied-configuration: | + {"apiVersion":"batch/v1","kind":"Job","metadata":{"annotations":{},"name":"check-span","namespace":"kuttl-test-amusing-sunbird"},"spec":{"backoffLimit":15,"template":{"spec":{"containers":[{"command":["./query"],"env":[{"name":"SERVICE_NAME","value":"smoke-test-service"},{"name":"QUERY_HOST","value":"https://agent-as-daemonset-query:443"},{"name":"SECRET_PATH","value":"/var/run/secrets/api-token/token"}],"image":"registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e","name":"asserts-container","volumeMounts":[{"mountPath":"/var/run/secrets/api-token","name":"token-api-volume"}]}],"restartPolicy":"OnFailure","volumes":[{"name":"token-api-volume","secret":{"secretName":"e2e-test"}}]}}}} + labels: + batch.kubernetes.io/controller-uid: dedc2ff1-c33d-4e16-af04-2abb8ee43c73 + batch.kubernetes.io/job-name: check-span + controller-uid: dedc2ff1-c33d-4e16-af04-2abb8ee43c73 + job-name: check-span + managedFields: + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:kubectl.kubernetes.io/last-applied-configuration: {} + f:spec: + f:backoffLimit: {} + f:completionMode: {} + f:completions: {} + f:parallelism: {} + f:suspend: {} + f:template: + f:spec: + f:containers: + k:{"name":"asserts-container"}: + .: {} + f:command: {} + f:env: + .: {} + k:{"name":"QUERY_HOST"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SECRET_PATH"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SERVICE_NAME"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:name: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/var/run/secrets/api-token"}: + .: {} + f:mountPath: {} + f:name: {} + f:dnsPolicy: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"token-api-volume"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + manager: kubectl-client-side-apply + operation: Update + time: "2023-11-27T14:32:22Z" + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:status: + f:active: {} + f:ready: {} + f:startTime: {} + f:uncountedTerminatedPods: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2023-11-27T14:40:41Z" name: check-span namespace: kuttl-test-amusing-sunbird +spec: + backoffLimit: 15 + completionMode: NonIndexed + completions: 1 + parallelism: 1 + selector: + matchLabels: + batch.kubernetes.io/controller-uid: dedc2ff1-c33d-4e16-af04-2abb8ee43c73 + suspend: false + template: + metadata: + creationTimestamp: null + labels: + batch.kubernetes.io/controller-uid: dedc2ff1-c33d-4e16-af04-2abb8ee43c73 + batch.kubernetes.io/job-name: check-span + controller-uid: dedc2ff1-c33d-4e16-af04-2abb8ee43c73 + job-name: check-span + spec: + containers: + - command: + - ./query + env: + - name: SERVICE_NAME + value: smoke-test-service + - name: QUERY_HOST + value: https://agent-as-daemonset-query:443 + - name: SECRET_PATH + value: /var/run/secrets/api-token/token + image: registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e + imagePullPolicy: IfNotPresent + name: asserts-container + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /var/run/secrets/api-token + name: token-api-volume + dnsPolicy: ClusterFirst + restartPolicy: OnFailure + schedulerName: default-scheduler + securityContext: {} + terminationGracePeriodSeconds: 30 + volumes: + - name: token-api-volume + secret: + defaultMode: 420 + secretName: e2e-test status: - succeeded: 1 + active: 1 + ready: 1 + startTime: "2023-11-27T14:32:22Z" + uncountedTerminatedPods: {} case.go:366: resource Job:kuttl-test-amusing-sunbird/check-span: .status.succeeded: key is missing from map logger.go:42: 14:42:22 | examples-agent-with-priority-class | examples-agent-with-priority-class events from ns kuttl-test-amusing-sunbird: logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:11 +0000 UTC Normal Pod agent-as-daemonset-54f7bc9697-q9dqb Binding Scheduled Successfully assigned kuttl-test-amusing-sunbird/agent-as-daemonset-54f7bc9697-q9dqb to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:11 +0000 UTC Normal Pod agent-as-daemonset-54f7bc9697-q9dqb AddedInterface Add eth0 [10.128.2.47/23] from ovn-kubernetes logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:11 +0000 UTC Normal Pod agent-as-daemonset-54f7bc9697-q9dqb.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:11 +0000 UTC Normal Pod agent-as-daemonset-54f7bc9697-q9dqb.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:11 +0000 UTC Normal Pod agent-as-daemonset-54f7bc9697-q9dqb.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:11 +0000 UTC Normal Pod agent-as-daemonset-54f7bc9697-q9dqb.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:11 +0000 UTC Normal Pod agent-as-daemonset-54f7bc9697-q9dqb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:11 +0000 UTC Normal Pod agent-as-daemonset-54f7bc9697-q9dqb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:11 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-54f7bc9697 SuccessfulCreate Created pod: agent-as-daemonset-54f7bc9697-q9dqb replicaset-controller logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:11 +0000 UTC Warning DaemonSet.apps agent-as-daemonset-agent-daemonset FailedCreate Error creating: pods "agent-as-daemonset-agent-daemonset-" is forbidden: unable to validate against any security context constraint: [provider "anyuid": Forbidden: not usable by user or serviceaccount, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5775: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5778: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6831: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6832: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 14271: Host ports are not allowed to be used, provider "restricted": Forbidden: not usable by user or serviceaccount, provider "nonroot-v2": Forbidden: not usable by user or serviceaccount, provider "nonroot": Forbidden: not usable by user or serviceaccount, provider "hostmount-anyuid": Forbidden: not usable by user or serviceaccount, provider "elasticsearch-scc": Forbidden: not usable by user or serviceaccount, provider "machine-api-termination-handler": Forbidden: not usable by user or serviceaccount, provider "daemonset-with-hostport": Forbidden: not usable by user or serviceaccount, provider "hostnetwork-v2": Forbidden: not usable by user or serviceaccount, provider "hostnetwork": Forbidden: not usable by user or serviceaccount, provider "hostaccess": Forbidden: not usable by user or serviceaccount, provider "node-exporter": Forbidden: not usable by user or serviceaccount, provider "privileged": Forbidden: not usable by user or serviceaccount] daemonset-controller logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:11 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-54f7bc9697 to 1 deployment-controller logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:22 +0000 UTC Normal Pod check-span-knff6 Binding Scheduled Successfully assigned kuttl-test-amusing-sunbird/check-span-knff6 to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:22 +0000 UTC Normal Pod check-span-knff6 AddedInterface Add eth0 [10.131.0.64/23] from ovn-kubernetes logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:22 +0000 UTC Normal Pod check-span-knff6.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:22 +0000 UTC Normal Pod check-span-knff6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:22 +0000 UTC Normal Pod check-span-knff6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:22 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-knff6 job-controller logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:22 +0000 UTC Normal Pod report-span-498wg Binding Scheduled Successfully assigned kuttl-test-amusing-sunbird/report-span-498wg to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:22 +0000 UTC Normal Pod report-span-498wg AddedInterface Add eth0 [10.129.2.92/23] from ovn-kubernetes logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:22 +0000 UTC Normal Pod report-span-498wg.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:22 +0000 UTC Normal Pod report-span-498wg.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:22 +0000 UTC Normal Pod report-span-498wg.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:32:22 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-498wg job-controller logger.go:42: 14:42:22 | examples-agent-with-priority-class | 2023-11-27 14:37:27 +0000 UTC Warning Pod check-span-knff6.spec.containers{asserts-container} BackOff Back-off restarting failed container asserts-container in pod check-span-knff6_kuttl-test-amusing-sunbird(c0133def-3207-4cd4-852f-8dd577352ed4) kubelet logger.go:42: 14:42:22 | examples-agent-with-priority-class | Deleting namespace: kuttl-test-amusing-sunbird === CONT kuttl/harness/examples-agent-as-daemonset logger.go:42: 14:42:34 | examples-agent-as-daemonset | Creating namespace: kuttl-test-sensible-swine logger.go:42: 14:42:34 | examples-agent-as-daemonset/0-install | starting test step 0-install logger.go:42: 14:42:34 | examples-agent-as-daemonset/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 14:42:34 | examples-agent-as-daemonset/0-install | ServiceAccount:kuttl-test-sensible-swine/jaeger-agent-daemonset created logger.go:42: 14:42:34 | examples-agent-as-daemonset/0-install | test step completed 0-install logger.go:42: 14:42:34 | examples-agent-as-daemonset/1-install | starting test step 1-install logger.go:42: 14:42:34 | examples-agent-as-daemonset/1-install | Jaeger:kuttl-test-sensible-swine/agent-as-daemonset created logger.go:42: 14:42:40 | examples-agent-as-daemonset/1-install | test step completed 1-install logger.go:42: 14:42:40 | examples-agent-as-daemonset/2-smoke-test | starting test step 2-smoke-test logger.go:42: 14:42:40 | examples-agent-as-daemonset/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 14:42:41 | examples-agent-as-daemonset/2-smoke-test | Warning: resource jaegers/agent-as-daemonset is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:42:48 | examples-agent-as-daemonset/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 14:42:48 | examples-agent-as-daemonset/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:42:48 | examples-agent-as-daemonset/2-smoke-test | job.batch/report-span created logger.go:42: 14:42:48 | examples-agent-as-daemonset/2-smoke-test | job.batch/check-span created logger.go:42: 14:42:59 | examples-agent-as-daemonset/2-smoke-test | test step completed 2-smoke-test logger.go:42: 14:42:59 | examples-agent-as-daemonset | examples-agent-as-daemonset events from ns kuttl-test-sensible-swine: logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:37 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-6f4dc94687 to 1 deployment-controller logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:38 +0000 UTC Normal Pod agent-as-daemonset-6f4dc94687-5nct8 Binding Scheduled Successfully assigned kuttl-test-sensible-swine/agent-as-daemonset-6f4dc94687-5nct8 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:38 +0000 UTC Normal Pod agent-as-daemonset-6f4dc94687-5nct8 AddedInterface Add eth0 [10.129.2.93/23] from ovn-kubernetes logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:38 +0000 UTC Normal Pod agent-as-daemonset-6f4dc94687-5nct8.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:38 +0000 UTC Normal Pod agent-as-daemonset-6f4dc94687-5nct8.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:38 +0000 UTC Normal Pod agent-as-daemonset-6f4dc94687-5nct8.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:38 +0000 UTC Normal Pod agent-as-daemonset-6f4dc94687-5nct8.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:38 +0000 UTC Normal Pod agent-as-daemonset-6f4dc94687-5nct8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:38 +0000 UTC Normal Pod agent-as-daemonset-6f4dc94687-5nct8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:38 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-6f4dc94687 SuccessfulCreate Created pod: agent-as-daemonset-6f4dc94687-5nct8 replicaset-controller logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:41 +0000 UTC Warning DaemonSet.apps agent-as-daemonset-agent-daemonset FailedCreate Error creating: pods "agent-as-daemonset-agent-daemonset-" is forbidden: unable to validate against any security context constraint: [provider "anyuid": Forbidden: not usable by user or serviceaccount, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5775: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5778: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6831: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6832: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 14271: Host ports are not allowed to be used, provider "restricted": Forbidden: not usable by user or serviceaccount, provider "nonroot-v2": Forbidden: not usable by user or serviceaccount, provider "nonroot": Forbidden: not usable by user or serviceaccount, provider "hostmount-anyuid": Forbidden: not usable by user or serviceaccount, provider "elasticsearch-scc": Forbidden: not usable by user or serviceaccount, provider "machine-api-termination-handler": Forbidden: not usable by user or serviceaccount, provider "daemonset-with-hostport": Forbidden: not usable by user or serviceaccount, provider "hostnetwork-v2": Forbidden: not usable by user or serviceaccount, provider "hostnetwork": Forbidden: not usable by user or serviceaccount, provider "hostaccess": Forbidden: not usable by user or serviceaccount, provider "node-exporter": Forbidden: not usable by user or serviceaccount, provider "privileged": Forbidden: not usable by user or serviceaccount] daemonset-controller logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:44 +0000 UTC Normal Pod agent-as-daemonset-6f4dc94687-5nct8.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:44 +0000 UTC Normal Pod agent-as-daemonset-6f4dc94687-5nct8.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:44 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-6f4dc94687 SuccessfulDelete Deleted pod: agent-as-daemonset-6f4dc94687-5nct8 replicaset-controller logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:44 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled down replica set agent-as-daemonset-6f4dc94687 to 0 from 1 deployment-controller logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:45 +0000 UTC Normal Pod agent-as-daemonset-5cf5b4777b-j5fqd Binding Scheduled Successfully assigned kuttl-test-sensible-swine/agent-as-daemonset-5cf5b4777b-j5fqd to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:45 +0000 UTC Normal Pod agent-as-daemonset-5cf5b4777b-j5fqd AddedInterface Add eth0 [10.128.2.48/23] from ovn-kubernetes logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:45 +0000 UTC Normal Pod agent-as-daemonset-5cf5b4777b-j5fqd.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:45 +0000 UTC Normal Pod agent-as-daemonset-5cf5b4777b-j5fqd.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:45 +0000 UTC Normal Pod agent-as-daemonset-5cf5b4777b-j5fqd.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:45 +0000 UTC Normal Pod agent-as-daemonset-5cf5b4777b-j5fqd.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:45 +0000 UTC Normal Pod agent-as-daemonset-5cf5b4777b-j5fqd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:45 +0000 UTC Normal Pod agent-as-daemonset-5cf5b4777b-j5fqd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:45 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-5cf5b4777b SuccessfulCreate Created pod: agent-as-daemonset-5cf5b4777b-j5fqd replicaset-controller logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:45 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-5cf5b4777b to 1 deployment-controller logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:48 +0000 UTC Normal Pod check-span-c8nqb Binding Scheduled Successfully assigned kuttl-test-sensible-swine/check-span-c8nqb to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:48 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-c8nqb job-controller logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:48 +0000 UTC Normal Pod report-span-s77rb Binding Scheduled Successfully assigned kuttl-test-sensible-swine/report-span-s77rb to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:48 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-s77rb job-controller logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:49 +0000 UTC Normal Pod check-span-c8nqb AddedInterface Add eth0 [10.131.0.65/23] from ovn-kubernetes logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:49 +0000 UTC Normal Pod check-span-c8nqb.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:49 +0000 UTC Normal Pod check-span-c8nqb.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:49 +0000 UTC Normal Pod check-span-c8nqb.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:49 +0000 UTC Normal Pod report-span-s77rb AddedInterface Add eth0 [10.129.2.94/23] from ovn-kubernetes logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:49 +0000 UTC Normal Pod report-span-s77rb.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:49 +0000 UTC Normal Pod report-span-s77rb.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:49 +0000 UTC Normal Pod report-span-s77rb.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:42:59 | examples-agent-as-daemonset | 2023-11-27 14:42:59 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:42:59 | examples-agent-as-daemonset | Deleting namespace: kuttl-test-sensible-swine === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (1486.66s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.75s) --- PASS: kuttl/harness/examples-service-types (53.83s) --- PASS: kuttl/harness/examples-with-sampling (59.55s) --- PASS: kuttl/harness/examples-with-cassandra (69.84s) --- PASS: kuttl/harness/examples-with-badger-and-volume (37.19s) --- PASS: kuttl/harness/examples-with-badger (37.09s) --- PASS: kuttl/harness/examples-simplest (37.89s) --- PASS: kuttl/harness/examples-simple-prod-with-volumes (67.90s) --- PASS: kuttl/harness/examples-simple-prod (69.53s) --- PASS: kuttl/harness/examples-business-application-injected-sidecar (41.22s) --- PASS: kuttl/harness/examples-openshift-with-htpasswd (22.73s) --- PASS: kuttl/harness/examples-openshift-agent-as-daemonset (62.96s) --- PASS: kuttl/harness/examples-collector-with-priority-class (32.87s) --- PASS: kuttl/harness/examples-all-in-one-with-options (38.20s) --- PASS: kuttl/harness/examples-auto-provision-kafka (185.62s) --- FAIL: kuttl/harness/examples-agent-with-priority-class (627.04s) --- PASS: kuttl/harness/examples-agent-as-daemonset (37.40s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name examples --report --output /logs/artifacts/examples.xml ./artifacts/kuttl-report.xml time="2023-11-27T14:43:12Z" level=debug msg="Setting a new name for the test suites" time="2023-11-27T14:43:12Z" level=debug msg="Removing 'artifacts' TestCase" time="2023-11-27T14:43:12Z" level=debug msg="normalizing test case names" time="2023-11-27T14:43:12Z" level=debug msg="examples/artifacts -> examples_artifacts" time="2023-11-27T14:43:12Z" level=debug msg="examples/examples-service-types -> examples_examples_service_types" time="2023-11-27T14:43:12Z" level=debug msg="examples/examples-with-sampling -> examples_examples_with_sampling" time="2023-11-27T14:43:12Z" level=debug msg="examples/examples-with-cassandra -> examples_examples_with_cassandra" time="2023-11-27T14:43:12Z" level=debug msg="examples/examples-with-badger-and-volume -> examples_examples_with_badger_and_volume" time="2023-11-27T14:43:12Z" level=debug msg="examples/examples-with-badger -> examples_examples_with_badger" time="2023-11-27T14:43:12Z" level=debug msg="examples/examples-simplest -> examples_examples_simplest" time="2023-11-27T14:43:12Z" level=debug msg="examples/examples-simple-prod-with-volumes -> examples_examples_simple_prod_with_volumes" time="2023-11-27T14:43:12Z" level=debug msg="examples/examples-simple-prod -> examples_examples_simple_prod" time="2023-11-27T14:43:12Z" level=debug msg="examples/examples-business-application-injected-sidecar -> examples_examples_business_application_injected_sidecar" time="2023-11-27T14:43:12Z" level=debug msg="examples/examples-openshift-with-htpasswd -> examples_examples_openshift_with_htpasswd" time="2023-11-27T14:43:12Z" level=debug msg="examples/examples-openshift-agent-as-daemonset -> examples_examples_openshift_agent_as_daemonset" time="2023-11-27T14:43:12Z" level=debug msg="examples/examples-collector-with-priority-class -> examples_examples_collector_with_priority_class" time="2023-11-27T14:43:12Z" level=debug msg="examples/examples-all-in-one-with-options -> examples_examples_all_in_one_with_options" time="2023-11-27T14:43:12Z" level=debug msg="examples/examples-auto-provision-kafka -> examples_examples_auto_provision_kafka" time="2023-11-27T14:43:12Z" level=debug msg="examples/examples-agent-with-priority-class -> examples_examples_agent_with_priority_class" time="2023-11-27T14:43:12Z" level=debug msg="examples/examples-agent-as-daemonset -> examples_examples_agent_as_daemonset" +---------------------------------------------------------+--------+ | NAME | RESULT | +---------------------------------------------------------+--------+ | examples_artifacts | passed | | examples_examples_service_types | passed | | examples_examples_with_sampling | passed | | examples_examples_with_cassandra | passed | | examples_examples_with_badger_and_volume | passed | | examples_examples_with_badger | passed | | examples_examples_simplest | passed | | examples_examples_simple_prod_with_volumes | passed | | examples_examples_simple_prod | passed | | examples_examples_business_application_injected_sidecar | passed | | examples_examples_openshift_with_htpasswd | passed | | examples_examples_openshift_agent_as_daemonset | passed | | examples_examples_collector_with_priority_class | passed | | examples_examples_all_in_one_with_options | passed | | examples_examples_auto_provision_kafka | passed | | examples_examples_agent_with_priority_class | failed | | examples_examples_agent_as_daemonset | passed | +---------------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh generate false true + '[' 3 -ne 3 ']' + test_suite_name=generate + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/generate.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-generate make[2]: Entering directory '/tmp/jaeger-tests' test -s /tmp/jaeger-tests/bin/operator-sdk || curl -sLo /tmp/jaeger-tests/bin/operator-sdk https://github.com/operator-framework/operator-sdk/releases/download/v1.27.0/operator-sdk_`go env GOOS`_`go env GOARCH` ./hack/install/install-golangci-lint.sh Installing golangci-lint golangci-lint 1.53.2 is installed already ./hack/install/install-goimports.sh Installing goimports Try 0... go install golang.org/x/tools/cmd/goimports@v0.1.12 >>>> Formatting code... ./.ci/format.sh >>>> Building... ./hack/install/install-dependencies.sh Installing go dependencies Try 0... go mod download GOOS= GOARCH= CGO_ENABLED=0 GO111MODULE=on go build -ldflags "-X "github.com/jaegertracing/jaeger-operator/pkg/version".version="1.51.0" -X "github.com/jaegertracing/jaeger-operator/pkg/version".buildDate=2023-11-27T14:43:14Z -X "github.com/jaegertracing/jaeger-operator/pkg/version".defaultJaeger="1.51.0"" -o "bin/jaeger-operator" main.go JAEGER_VERSION="1.51.0" ./tests/e2e/generate/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-11-25-110147 True False 47m Cluster version is 4.15.0-0.nightly-2023-11-25-110147' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-11-25-110147 True False 47m Cluster version is 4.15.0-0.nightly-2023-11-25-110147' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z '' ']' ++ KAFKA_USE_CUSTOM_PODSET=false ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/generate/render.sh ++ export SUITE_DIR=./tests/e2e/generate ++ SUITE_DIR=./tests/e2e/generate ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/generate ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test generate 'This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 2 -ne 2 ']' + test_name=generate + message='This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/generate/_build + '[' _build '!=' _build ']' + rm -rf generate + warning 'generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed\e[0m' WAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running generate E2E tests' Running generate E2E tests + cd tests/e2e/generate/_build + set +e + KUBECONFIG=/tmp/kubeconfig-4018005699 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 14:43:31 | artifacts | Creating namespace: kuttl-test-assured-trout logger.go:42: 14:43:31 | artifacts | artifacts events from ns kuttl-test-assured-trout: logger.go:42: 14:43:31 | artifacts | Deleting namespace: kuttl-test-assured-trout === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (5.70s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.65s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name generate --report --output /logs/artifacts/generate.xml ./artifacts/kuttl-report.xml time="2023-11-27T14:43:36Z" level=debug msg="Setting a new name for the test suites" time="2023-11-27T14:43:36Z" level=debug msg="Removing 'artifacts' TestCase" time="2023-11-27T14:43:36Z" level=debug msg="normalizing test case names" time="2023-11-27T14:43:36Z" level=debug msg="generate/artifacts -> generate_artifacts" +--------------------+--------+ | NAME | RESULT | +--------------------+--------+ | generate_artifacts | passed | +--------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh miscellaneous false true + '[' 3 -ne 3 ']' + test_suite_name=miscellaneous + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/miscellaneous.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-miscellaneous make[2]: Entering directory '/tmp/jaeger-tests' SKIP_ES_EXTERNAL=true ./tests/e2e/miscellaneous/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-11-25-110147 True False 47m Cluster version is 4.15.0-0.nightly-2023-11-25-110147' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-11-25-110147 True False 47m Cluster version is 4.15.0-0.nightly-2023-11-25-110147' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z '' ']' ++ KAFKA_USE_CUSTOM_PODSET=false ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/miscellaneous/render.sh ++ export SUITE_DIR=./tests/e2e/miscellaneous ++ SUITE_DIR=./tests/e2e/miscellaneous ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/miscellaneous ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test cassandra-spark 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=cassandra-spark + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf cassandra-spark + warning 'cassandra-spark: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: cassandra-spark: Test not supported in OpenShift\e[0m' WAR: cassandra-spark: Test not supported in OpenShift + start_test collector-autoscale + '[' 1 -ne 1 ']' + test_name=collector-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-autoscale\e[0m' Rendering files for test collector-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p collector-autoscale + cd collector-autoscale + jaeger_name=simple-prod + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + ELASTICSEARCH_NODECOUNT=1 + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.resources.requests.memory="200m"' 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.autoscale=true 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.minReplicas=1 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.maxReplicas=2 01-install.yaml + kubectl api-versions + grep autoscaling/v2beta2 -q + rm ./04-assert.yaml + generate_otlp_e2e_tests http + test_protocol=http + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-http\e[0m' Rendering files for test collector-otlp-allinone-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-autoscale + '[' collector-autoscale '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-http + cd collector-otlp-allinone-http + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger http true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-http\e[0m' Rendering files for test collector-otlp-production-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-http + '[' collector-otlp-allinone-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-http + cd collector-otlp-production-http + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger http true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + generate_otlp_e2e_tests grpc + test_protocol=grpc + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-grpc\e[0m' Rendering files for test collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-http + '[' collector-otlp-production-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-grpc + cd collector-otlp-allinone-grpc + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-grpc\e[0m' Rendering files for test collector-otlp-production-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-grpc + '[' collector-otlp-allinone-grpc '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-grpc + cd collector-otlp-production-grpc + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + '[' true = true ']' + skip_test istio 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=istio + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-grpc + '[' collector-otlp-production-grpc '!=' _build ']' + cd .. + rm -rf istio + warning 'istio: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: istio: Test not supported in OpenShift\e[0m' WAR: istio: Test not supported in OpenShift + '[' true = true ']' + skip_test outside-cluster 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=outside-cluster + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf outside-cluster + warning 'outside-cluster: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: outside-cluster: Test not supported in OpenShift\e[0m' WAR: outside-cluster: Test not supported in OpenShift + start_test set-custom-img + '[' 1 -ne 1 ']' + test_name=set-custom-img + echo =========================================================================== =========================================================================== + info 'Rendering files for test set-custom-img' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test set-custom-img\e[0m' Rendering files for test set-custom-img + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p set-custom-img + cd set-custom-img + jaeger_name=my-jaeger + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.image="test"' ./02-install.yaml + '[' true = true ']' + skip_test non-cluster-wide 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=non-cluster-wide + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/set-custom-img + '[' set-custom-img '!=' _build ']' + cd .. + rm -rf non-cluster-wide + warning 'non-cluster-wide: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: non-cluster-wide: Test not supported in OpenShift\e[0m' WAR: non-cluster-wide: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running miscellaneous E2E tests' Running miscellaneous E2E tests + cd tests/e2e/miscellaneous/_build + set +e + KUBECONFIG=/tmp/kubeconfig-4018005699 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/collector-autoscale === PAUSE kuttl/harness/collector-autoscale === RUN kuttl/harness/collector-otlp-allinone-grpc === PAUSE kuttl/harness/collector-otlp-allinone-grpc === RUN kuttl/harness/collector-otlp-allinone-http === PAUSE kuttl/harness/collector-otlp-allinone-http === RUN kuttl/harness/collector-otlp-production-grpc === PAUSE kuttl/harness/collector-otlp-production-grpc === RUN kuttl/harness/collector-otlp-production-http === PAUSE kuttl/harness/collector-otlp-production-http === RUN kuttl/harness/set-custom-img === PAUSE kuttl/harness/set-custom-img === CONT kuttl/harness/artifacts logger.go:42: 14:43:47 | artifacts | Creating namespace: kuttl-test-peaceful-leech logger.go:42: 14:43:47 | artifacts | artifacts events from ns kuttl-test-peaceful-leech: logger.go:42: 14:43:47 | artifacts | Deleting namespace: kuttl-test-peaceful-leech === CONT kuttl/harness/collector-otlp-production-grpc logger.go:42: 14:43:53 | collector-otlp-production-grpc | Creating namespace: kuttl-test-gentle-honeybee logger.go:42: 14:43:53 | collector-otlp-production-grpc/1-install | starting test step 1-install logger.go:42: 14:43:53 | collector-otlp-production-grpc/1-install | Jaeger:kuttl-test-gentle-honeybee/my-jaeger created logger.go:42: 14:44:29 | collector-otlp-production-grpc/1-install | test step completed 1-install logger.go:42: 14:44:29 | collector-otlp-production-grpc/2-smoke-test | starting test step 2-smoke-test logger.go:42: 14:44:29 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 14:44:30 | collector-otlp-production-grpc/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:44:37 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 14:44:37 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:44:37 | collector-otlp-production-grpc/2-smoke-test | job.batch/report-span created logger.go:42: 14:44:37 | collector-otlp-production-grpc/2-smoke-test | job.batch/check-span created logger.go:42: 14:44:56 | collector-otlp-production-grpc/2-smoke-test | test step completed 2-smoke-test logger.go:42: 14:44:56 | collector-otlp-production-grpc | collector-otlp-production-grpc events from ns kuttl-test-gentle-honeybee: logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:43:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgentlehoneybeemyjaeger-1-85585d4d4w8 Binding Scheduled Successfully assigned kuttl-test-gentle-honeybee/elasticsearch-cdm-kuttltestgentlehoneybeemyjaeger-1-85585d4d4w8 to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:43:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgentlehoneybeemyjaeger-1-85585d4d4w8 AddedInterface Add eth0 [10.128.2.49/23] from ovn-kubernetes logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:43:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgentlehoneybeemyjaeger-1-85585d4d4w8.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:56ea62bfb0ca36e19a7b21aff3676e49511f05f72da5e76d6427fd8240f328a8" already present on machine kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:43:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgentlehoneybeemyjaeger-1-85585d4d4w8.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:43:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgentlehoneybeemyjaeger-1-85585d4d4w8.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:43:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgentlehoneybeemyjaeger-1-85585d4d4w8.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:79427bea6b5c37894b9782c3821d8b9074838e606daa4a743b2ae060856fa98a" already present on machine kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:43:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgentlehoneybeemyjaeger-1-85585d4d4w8.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:43:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgentlehoneybeemyjaeger-1-85585d4d4w8.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:43:59 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestgentlehoneybeemyjaeger-1-85585dd474 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestgentlehoneybeemyjaeger-1-85585d4d4w8 replicaset-controller logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:43:59 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestgentlehoneybeemyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestgentlehoneybeemyjaeger-1-85585dd474 to 1 deployment-controller logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:09 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestgentlehoneybeemyjaeger-1-85585d4d4w8.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:14 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestgentlehoneybeemyjaeger-1-85585d4d4w8.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:26 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-zk2pz Binding Scheduled Successfully assigned kuttl-test-gentle-honeybee/my-jaeger-collector-5489f5bd9b-zk2pz to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:26 +0000 UTC Warning Pod my-jaeger-collector-5489f5bd9b-zk2pz FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:26 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5489f5bd9b SuccessfulCreate Created pod: my-jaeger-collector-5489f5bd9b-zk2pz replicaset-controller logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:26 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-5489f5bd9b to 1 deployment-controller logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:26 +0000 UTC Normal Pod my-jaeger-query-74d6f5d8cd-p54c8 Binding Scheduled Successfully assigned kuttl-test-gentle-honeybee/my-jaeger-query-74d6f5d8cd-p54c8 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:26 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-74d6f5d8cd SuccessfulCreate Created pod: my-jaeger-query-74d6f5d8cd-p54c8 replicaset-controller logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:26 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-74d6f5d8cd to 1 deployment-controller logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:27 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-zk2pz AddedInterface Add eth0 [10.129.2.95/23] from ovn-kubernetes logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:27 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-zk2pz.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" already present on machine kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:27 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-zk2pz.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:27 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-zk2pz.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:27 +0000 UTC Warning Pod my-jaeger-query-74d6f5d8cd-p54c8 FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-oauth-proxy-tls" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:27 +0000 UTC Warning Pod my-jaeger-query-74d6f5d8cd-p54c8 FailedMount MountVolume.SetUp failed for volume "my-jaeger-service-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:28 +0000 UTC Normal Pod my-jaeger-query-74d6f5d8cd-p54c8 AddedInterface Add eth0 [10.129.2.96/23] from ovn-kubernetes logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:28 +0000 UTC Normal Pod my-jaeger-query-74d6f5d8cd-p54c8.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:28 +0000 UTC Normal Pod my-jaeger-query-74d6f5d8cd-p54c8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:28 +0000 UTC Normal Pod my-jaeger-query-74d6f5d8cd-p54c8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:28 +0000 UTC Normal Pod my-jaeger-query-74d6f5d8cd-p54c8.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:28 +0000 UTC Normal Pod my-jaeger-query-74d6f5d8cd-p54c8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:28 +0000 UTC Normal Pod my-jaeger-query-74d6f5d8cd-p54c8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:28 +0000 UTC Normal Pod my-jaeger-query-74d6f5d8cd-p54c8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:28 +0000 UTC Normal Pod my-jaeger-query-74d6f5d8cd-p54c8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:28 +0000 UTC Normal Pod my-jaeger-query-74d6f5d8cd-p54c8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:33 +0000 UTC Normal Pod my-jaeger-query-74d6f5d8cd-p54c8.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:33 +0000 UTC Normal Pod my-jaeger-query-74d6f5d8cd-p54c8.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:33 +0000 UTC Normal Pod my-jaeger-query-74d6f5d8cd-p54c8.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:33 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-74d6f5d8cd SuccessfulDelete Deleted pod: my-jaeger-query-74d6f5d8cd-p54c8 replicaset-controller logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:33 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-74d6f5d8cd to 0 from 1 deployment-controller logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:34 +0000 UTC Normal Pod my-jaeger-query-dd494d747-4zmgx Binding Scheduled Successfully assigned kuttl-test-gentle-honeybee/my-jaeger-query-dd494d747-4zmgx to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:34 +0000 UTC Normal Pod my-jaeger-query-dd494d747-4zmgx AddedInterface Add eth0 [10.131.0.66/23] from ovn-kubernetes logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:34 +0000 UTC Normal Pod my-jaeger-query-dd494d747-4zmgx.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:34 +0000 UTC Normal Pod my-jaeger-query-dd494d747-4zmgx.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:34 +0000 UTC Normal Pod my-jaeger-query-dd494d747-4zmgx.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:34 +0000 UTC Normal Pod my-jaeger-query-dd494d747-4zmgx.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:34 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-dd494d747 SuccessfulCreate Created pod: my-jaeger-query-dd494d747-4zmgx replicaset-controller logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:34 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-dd494d747 to 1 deployment-controller logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:35 +0000 UTC Normal Pod my-jaeger-query-dd494d747-4zmgx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:35 +0000 UTC Normal Pod my-jaeger-query-dd494d747-4zmgx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:35 +0000 UTC Normal Pod my-jaeger-query-dd494d747-4zmgx.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:35 +0000 UTC Normal Pod my-jaeger-query-dd494d747-4zmgx.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:35 +0000 UTC Normal Pod my-jaeger-query-dd494d747-4zmgx.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:37 +0000 UTC Normal Pod check-span-hh9jl Binding Scheduled Successfully assigned kuttl-test-gentle-honeybee/check-span-hh9jl to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:37 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-hh9jl job-controller logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:37 +0000 UTC Normal Pod report-span-vkm26 Binding Scheduled Successfully assigned kuttl-test-gentle-honeybee/report-span-vkm26 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:37 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-vkm26 job-controller logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:38 +0000 UTC Normal Pod check-span-hh9jl AddedInterface Add eth0 [10.129.2.98/23] from ovn-kubernetes logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:38 +0000 UTC Normal Pod check-span-hh9jl.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:38 +0000 UTC Normal Pod check-span-hh9jl.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:38 +0000 UTC Normal Pod check-span-hh9jl.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:38 +0000 UTC Normal Pod report-span-vkm26 AddedInterface Add eth0 [10.129.2.97/23] from ovn-kubernetes logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:38 +0000 UTC Normal Pod report-span-vkm26.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:38 +0000 UTC Normal Pod report-span-vkm26.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:38 +0000 UTC Normal Pod report-span-vkm26.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:44:56 | collector-otlp-production-grpc | 2023-11-27 14:44:56 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:44:56 | collector-otlp-production-grpc | Deleting namespace: kuttl-test-gentle-honeybee === CONT kuttl/harness/set-custom-img logger.go:42: 14:45:08 | set-custom-img | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:45:08 | set-custom-img | Ignoring check-collector-img.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:45:08 | set-custom-img | Creating namespace: kuttl-test-becoming-spider logger.go:42: 14:45:08 | set-custom-img/1-install | starting test step 1-install logger.go:42: 14:45:08 | set-custom-img/1-install | Jaeger:kuttl-test-becoming-spider/my-jaeger created logger.go:42: 14:45:45 | set-custom-img/1-install | test step completed 1-install logger.go:42: 14:45:45 | set-custom-img/2-install | starting test step 2-install logger.go:42: 14:45:45 | set-custom-img/2-install | Jaeger:kuttl-test-becoming-spider/my-jaeger updated logger.go:42: 14:45:45 | set-custom-img/2-install | test step completed 2-install logger.go:42: 14:45:45 | set-custom-img/3-check-image | starting test step 3-check-image logger.go:42: 14:45:45 | set-custom-img/3-check-image | running command: [sh -c ./check-collector-img.sh] logger.go:42: 14:45:45 | set-custom-img/3-check-image | Collector image missmatch. Expected: test. Has: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856 logger.go:42: 14:45:50 | set-custom-img/3-check-image | Collector image asserted properly! logger.go:42: 14:45:50 | set-custom-img/3-check-image | test step completed 3-check-image logger.go:42: 14:45:50 | set-custom-img | set-custom-img events from ns kuttl-test-becoming-spider: logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:15 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestbecomingspidermyjaeger-1-5dd9b577d8 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestbecomingspidermyjaeger-1-5dd9b5tmp75 replicaset-controller logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingspidermyjaeger-1-5dd9b5tmp75 Binding Scheduled Successfully assigned kuttl-test-becoming-spider/elasticsearch-cdm-kuttltestbecomingspidermyjaeger-1-5dd9b5tmp75 to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingspidermyjaeger-1-5dd9b5tmp75 AddedInterface Add eth0 [10.128.2.50/23] from ovn-kubernetes logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:15 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestbecomingspidermyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestbecomingspidermyjaeger-1-5dd9b577d8 to 1 deployment-controller logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingspidermyjaeger-1-5dd9b5tmp75.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:56ea62bfb0ca36e19a7b21aff3676e49511f05f72da5e76d6427fd8240f328a8" already present on machine kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingspidermyjaeger-1-5dd9b5tmp75.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingspidermyjaeger-1-5dd9b5tmp75.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingspidermyjaeger-1-5dd9b5tmp75.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:79427bea6b5c37894b9782c3821d8b9074838e606daa4a743b2ae060856fa98a" already present on machine kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingspidermyjaeger-1-5dd9b5tmp75.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbecomingspidermyjaeger-1-5dd9b5tmp75.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:31 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestbecomingspidermyjaeger-1-5dd9b5tmp75.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:42 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-blknz Binding Scheduled Successfully assigned kuttl-test-becoming-spider/my-jaeger-collector-5489f5bd9b-blknz to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:42 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5489f5bd9b SuccessfulCreate Created pod: my-jaeger-collector-5489f5bd9b-blknz replicaset-controller logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:42 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-5489f5bd9b to 1 deployment-controller logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:42 +0000 UTC Normal Pod my-jaeger-query-7b4fdb8984-pf6sr Binding Scheduled Successfully assigned kuttl-test-becoming-spider/my-jaeger-query-7b4fdb8984-pf6sr to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:42 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7b4fdb8984 SuccessfulCreate Created pod: my-jaeger-query-7b4fdb8984-pf6sr replicaset-controller logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:42 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7b4fdb8984 to 1 deployment-controller logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:43 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-blknz AddedInterface Add eth0 [10.131.0.67/23] from ovn-kubernetes logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:43 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-blknz.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" already present on machine kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:43 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-blknz.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:43 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-blknz.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:43 +0000 UTC Normal Pod my-jaeger-query-7b4fdb8984-pf6sr AddedInterface Add eth0 [10.129.2.100/23] from ovn-kubernetes logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:43 +0000 UTC Normal Pod my-jaeger-query-7b4fdb8984-pf6sr.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:43 +0000 UTC Normal Pod my-jaeger-query-7b4fdb8984-pf6sr.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:43 +0000 UTC Normal Pod my-jaeger-query-7b4fdb8984-pf6sr.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:43 +0000 UTC Normal Pod my-jaeger-query-7b4fdb8984-pf6sr.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:43 +0000 UTC Normal Pod my-jaeger-query-7b4fdb8984-pf6sr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:43 +0000 UTC Normal Pod my-jaeger-query-7b4fdb8984-pf6sr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:43 +0000 UTC Normal Pod my-jaeger-query-7b4fdb8984-pf6sr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:43 +0000 UTC Normal Pod my-jaeger-query-7b4fdb8984-pf6sr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:43 +0000 UTC Normal Pod my-jaeger-query-7b4fdb8984-pf6sr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:46 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5489f5bd9b SuccessfulDelete Deleted pod: my-jaeger-collector-5489f5bd9b-blknz replicaset-controller logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:46 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-5489f5bd9b to 0 from 1 deployment-controller logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:47 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-blknz.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:48 +0000 UTC Normal Pod my-jaeger-collector-7fd96ccd44-bm6rf Binding Scheduled Successfully assigned kuttl-test-becoming-spider/my-jaeger-collector-7fd96ccd44-bm6rf to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:48 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7fd96ccd44 SuccessfulCreate Created pod: my-jaeger-collector-7fd96ccd44-bm6rf replicaset-controller logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:48 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7fd96ccd44 to 1 deployment-controller logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:49 +0000 UTC Normal Pod my-jaeger-collector-7fd96ccd44-bm6rf AddedInterface Add eth0 [10.131.0.68/23] from ovn-kubernetes logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:49 +0000 UTC Normal Pod my-jaeger-collector-7fd96ccd44-bm6rf.spec.containers{jaeger-collector} Pulling Pulling image "test" kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:49 +0000 UTC Warning Pod my-jaeger-collector-7fd96ccd44-bm6rf.spec.containers{jaeger-collector} Failed Failed to pull image "test": reading manifest latest in docker.io/library/test: requested access to the resource is denied kubelet logger.go:42: 14:45:50 | set-custom-img | 2023-11-27 14:45:49 +0000 UTC Warning Pod my-jaeger-collector-7fd96ccd44-bm6rf.spec.containers{jaeger-collector} Failed Error: ErrImagePull kubelet logger.go:42: 14:45:50 | set-custom-img | Deleting namespace: kuttl-test-becoming-spider === CONT kuttl/harness/collector-otlp-production-http logger.go:42: 14:45:56 | collector-otlp-production-http | Creating namespace: kuttl-test-enormous-falcon logger.go:42: 14:45:56 | collector-otlp-production-http/1-install | starting test step 1-install logger.go:42: 14:45:56 | collector-otlp-production-http/1-install | Jaeger:kuttl-test-enormous-falcon/my-jaeger created logger.go:42: 14:46:32 | collector-otlp-production-http/1-install | test step completed 1-install logger.go:42: 14:46:32 | collector-otlp-production-http/2-smoke-test | starting test step 2-smoke-test logger.go:42: 14:46:32 | collector-otlp-production-http/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 14:46:33 | collector-otlp-production-http/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:46:39 | collector-otlp-production-http/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 14:46:39 | collector-otlp-production-http/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:46:40 | collector-otlp-production-http/2-smoke-test | job.batch/report-span created logger.go:42: 14:46:40 | collector-otlp-production-http/2-smoke-test | job.batch/check-span created logger.go:42: 14:46:52 | collector-otlp-production-http/2-smoke-test | test step completed 2-smoke-test logger.go:42: 14:46:52 | collector-otlp-production-http | collector-otlp-production-http events from ns kuttl-test-enormous-falcon: logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:02 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestenormousfalconmyjaeger-1-85997f74fb SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestenormousfalconmyjaeger-1-85997fvln5m replicaset-controller logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestenormousfalconmyjaeger-1-85997fvln5m Binding Scheduled Successfully assigned kuttl-test-enormous-falcon/elasticsearch-cdm-kuttltestenormousfalconmyjaeger-1-85997fvln5m to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:02 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestenormousfalconmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestenormousfalconmyjaeger-1-85997f74fb to 1 deployment-controller logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestenormousfalconmyjaeger-1-85997fvln5m AddedInterface Add eth0 [10.128.2.51/23] from ovn-kubernetes logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestenormousfalconmyjaeger-1-85997fvln5m.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:56ea62bfb0ca36e19a7b21aff3676e49511f05f72da5e76d6427fd8240f328a8" already present on machine kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestenormousfalconmyjaeger-1-85997fvln5m.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestenormousfalconmyjaeger-1-85997fvln5m.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestenormousfalconmyjaeger-1-85997fvln5m.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:79427bea6b5c37894b9782c3821d8b9074838e606daa4a743b2ae060856fa98a" already present on machine kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestenormousfalconmyjaeger-1-85997fvln5m.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestenormousfalconmyjaeger-1-85997fvln5m.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:13 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestenormousfalconmyjaeger-1-85997fvln5m.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:18 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestenormousfalconmyjaeger-1-85997fvln5m.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:29 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-x9fdg Binding Scheduled Successfully assigned kuttl-test-enormous-falcon/my-jaeger-collector-5489f5bd9b-x9fdg to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:29 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5489f5bd9b SuccessfulCreate Created pod: my-jaeger-collector-5489f5bd9b-x9fdg replicaset-controller logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:29 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-5489f5bd9b to 1 deployment-controller logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:29 +0000 UTC Normal Pod my-jaeger-query-fbb876584-fbpws Binding Scheduled Successfully assigned kuttl-test-enormous-falcon/my-jaeger-query-fbb876584-fbpws to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:29 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-fbb876584 SuccessfulCreate Created pod: my-jaeger-query-fbb876584-fbpws replicaset-controller logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:29 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-fbb876584 to 1 deployment-controller logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:30 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-x9fdg AddedInterface Add eth0 [10.129.2.101/23] from ovn-kubernetes logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:30 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-x9fdg.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" already present on machine kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:30 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-x9fdg.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:30 +0000 UTC Normal Pod my-jaeger-collector-5489f5bd9b-x9fdg.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:30 +0000 UTC Normal Pod my-jaeger-query-fbb876584-fbpws AddedInterface Add eth0 [10.129.2.102/23] from ovn-kubernetes logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:30 +0000 UTC Normal Pod my-jaeger-query-fbb876584-fbpws.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:30 +0000 UTC Normal Pod my-jaeger-query-fbb876584-fbpws.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:30 +0000 UTC Normal Pod my-jaeger-query-fbb876584-fbpws.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:30 +0000 UTC Normal Pod my-jaeger-query-fbb876584-fbpws.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:30 +0000 UTC Normal Pod my-jaeger-query-fbb876584-fbpws.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:30 +0000 UTC Normal Pod my-jaeger-query-fbb876584-fbpws.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:30 +0000 UTC Normal Pod my-jaeger-query-fbb876584-fbpws.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:30 +0000 UTC Normal Pod my-jaeger-query-fbb876584-fbpws.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:30 +0000 UTC Normal Pod my-jaeger-query-fbb876584-fbpws.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:34 +0000 UTC Normal Pod my-jaeger-query-fbb876584-fbpws.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:34 +0000 UTC Normal Pod my-jaeger-query-fbb876584-fbpws.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:34 +0000 UTC Normal Pod my-jaeger-query-fbb876584-fbpws.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:34 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-fbb876584 SuccessfulDelete Deleted pod: my-jaeger-query-fbb876584-fbpws replicaset-controller logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:34 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-fbb876584 to 0 from 1 deployment-controller logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:35 +0000 UTC Normal Pod my-jaeger-query-5455fcd848-pwlvd Binding Scheduled Successfully assigned kuttl-test-enormous-falcon/my-jaeger-query-5455fcd848-pwlvd to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:35 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5455fcd848 SuccessfulCreate Created pod: my-jaeger-query-5455fcd848-pwlvd replicaset-controller logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:35 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-5455fcd848 to 1 deployment-controller logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:36 +0000 UTC Normal Pod my-jaeger-query-5455fcd848-pwlvd AddedInterface Add eth0 [10.131.0.69/23] from ovn-kubernetes logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:36 +0000 UTC Normal Pod my-jaeger-query-5455fcd848-pwlvd.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:36 +0000 UTC Normal Pod my-jaeger-query-5455fcd848-pwlvd.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:36 +0000 UTC Normal Pod my-jaeger-query-5455fcd848-pwlvd.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:36 +0000 UTC Normal Pod my-jaeger-query-5455fcd848-pwlvd.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:36 +0000 UTC Normal Pod my-jaeger-query-5455fcd848-pwlvd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:36 +0000 UTC Normal Pod my-jaeger-query-5455fcd848-pwlvd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:36 +0000 UTC Normal Pod my-jaeger-query-5455fcd848-pwlvd.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:36 +0000 UTC Normal Pod my-jaeger-query-5455fcd848-pwlvd.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:36 +0000 UTC Normal Pod my-jaeger-query-5455fcd848-pwlvd.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:40 +0000 UTC Normal Pod check-span-r8cdg Binding Scheduled Successfully assigned kuttl-test-enormous-falcon/check-span-r8cdg to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:40 +0000 UTC Normal Pod check-span-r8cdg AddedInterface Add eth0 [10.129.2.104/23] from ovn-kubernetes logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:40 +0000 UTC Normal Pod check-span-r8cdg.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:40 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-r8cdg job-controller logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:40 +0000 UTC Normal Pod report-span-wdhxt Binding Scheduled Successfully assigned kuttl-test-enormous-falcon/report-span-wdhxt to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:40 +0000 UTC Normal Pod report-span-wdhxt AddedInterface Add eth0 [10.129.2.103/23] from ovn-kubernetes logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:40 +0000 UTC Normal Pod report-span-wdhxt.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:40 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-wdhxt job-controller logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:41 +0000 UTC Normal Pod check-span-r8cdg.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:41 +0000 UTC Normal Pod check-span-r8cdg.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:41 +0000 UTC Normal Pod report-span-wdhxt.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:41 +0000 UTC Normal Pod report-span-wdhxt.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:44 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:46:52 | collector-otlp-production-http | 2023-11-27 14:46:51 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:46:52 | collector-otlp-production-http | Deleting namespace: kuttl-test-enormous-falcon === CONT kuttl/harness/collector-otlp-allinone-grpc logger.go:42: 14:47:04 | collector-otlp-allinone-grpc | Creating namespace: kuttl-test-casual-lion logger.go:42: 14:47:04 | collector-otlp-allinone-grpc/0-install | starting test step 0-install logger.go:42: 14:47:04 | collector-otlp-allinone-grpc/0-install | Jaeger:kuttl-test-casual-lion/my-jaeger created logger.go:42: 14:47:10 | collector-otlp-allinone-grpc/0-install | test step completed 0-install logger.go:42: 14:47:10 | collector-otlp-allinone-grpc/1-smoke-test | starting test step 1-smoke-test logger.go:42: 14:47:10 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 14:47:11 | collector-otlp-allinone-grpc/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:47:17 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 14:47:18 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:47:18 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/report-span created logger.go:42: 14:47:18 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/check-span created logger.go:42: 14:47:37 | collector-otlp-allinone-grpc/1-smoke-test | test step completed 1-smoke-test logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | collector-otlp-allinone-grpc events from ns kuttl-test-casual-lion: logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:07 +0000 UTC Normal Pod my-jaeger-575c6d6c7b-sjvkr Binding Scheduled Successfully assigned kuttl-test-casual-lion/my-jaeger-575c6d6c7b-sjvkr to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:07 +0000 UTC Normal ReplicaSet.apps my-jaeger-575c6d6c7b SuccessfulCreate Created pod: my-jaeger-575c6d6c7b-sjvkr replicaset-controller logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:07 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-575c6d6c7b to 1 deployment-controller logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:08 +0000 UTC Normal Pod my-jaeger-575c6d6c7b-sjvkr AddedInterface Add eth0 [10.128.2.52/23] from ovn-kubernetes logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:08 +0000 UTC Normal Pod my-jaeger-575c6d6c7b-sjvkr.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:08 +0000 UTC Normal Pod my-jaeger-575c6d6c7b-sjvkr.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:08 +0000 UTC Normal Pod my-jaeger-575c6d6c7b-sjvkr.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:08 +0000 UTC Normal Pod my-jaeger-575c6d6c7b-sjvkr.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:08 +0000 UTC Normal Pod my-jaeger-575c6d6c7b-sjvkr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:08 +0000 UTC Normal Pod my-jaeger-575c6d6c7b-sjvkr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:14 +0000 UTC Normal Pod my-jaeger-575c6d6c7b-sjvkr.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:14 +0000 UTC Normal Pod my-jaeger-575c6d6c7b-sjvkr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:14 +0000 UTC Normal ReplicaSet.apps my-jaeger-575c6d6c7b SuccessfulDelete Deleted pod: my-jaeger-575c6d6c7b-sjvkr replicaset-controller logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:14 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-575c6d6c7b to 0 from 1 deployment-controller logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:15 +0000 UTC Normal Pod my-jaeger-55d474b697-k5vl6 Binding Scheduled Successfully assigned kuttl-test-casual-lion/my-jaeger-55d474b697-k5vl6 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:15 +0000 UTC Normal Pod my-jaeger-55d474b697-k5vl6 AddedInterface Add eth0 [10.129.2.105/23] from ovn-kubernetes logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:15 +0000 UTC Normal Pod my-jaeger-55d474b697-k5vl6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:15 +0000 UTC Normal Pod my-jaeger-55d474b697-k5vl6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:15 +0000 UTC Normal Pod my-jaeger-55d474b697-k5vl6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:15 +0000 UTC Normal Pod my-jaeger-55d474b697-k5vl6.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:15 +0000 UTC Normal Pod my-jaeger-55d474b697-k5vl6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:15 +0000 UTC Normal Pod my-jaeger-55d474b697-k5vl6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:15 +0000 UTC Normal ReplicaSet.apps my-jaeger-55d474b697 SuccessfulCreate Created pod: my-jaeger-55d474b697-k5vl6 replicaset-controller logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:15 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-55d474b697 to 1 deployment-controller logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:18 +0000 UTC Normal Pod check-span-fcq6w Binding Scheduled Successfully assigned kuttl-test-casual-lion/check-span-fcq6w to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:18 +0000 UTC Normal Pod check-span-fcq6w AddedInterface Add eth0 [10.131.0.70/23] from ovn-kubernetes logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:18 +0000 UTC Normal Pod check-span-fcq6w.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:18 +0000 UTC Normal Pod check-span-fcq6w.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:18 +0000 UTC Normal Pod check-span-fcq6w.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:18 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-fcq6w job-controller logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:18 +0000 UTC Normal Pod report-span-zwbnm Binding Scheduled Successfully assigned kuttl-test-casual-lion/report-span-zwbnm to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:18 +0000 UTC Normal Pod report-span-zwbnm AddedInterface Add eth0 [10.128.2.53/23] from ovn-kubernetes logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:18 +0000 UTC Normal Pod report-span-zwbnm.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:18 +0000 UTC Normal Pod report-span-zwbnm.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:18 +0000 UTC Normal Pod report-span-zwbnm.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:18 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-zwbnm job-controller logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | 2023-11-27 14:47:37 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:47:37 | collector-otlp-allinone-grpc | Deleting namespace: kuttl-test-casual-lion === CONT kuttl/harness/collector-otlp-allinone-http logger.go:42: 14:47:49 | collector-otlp-allinone-http | Creating namespace: kuttl-test-grateful-seal logger.go:42: 14:47:49 | collector-otlp-allinone-http/0-install | starting test step 0-install logger.go:42: 14:47:49 | collector-otlp-allinone-http/0-install | Jaeger:kuttl-test-grateful-seal/my-jaeger created logger.go:42: 14:47:55 | collector-otlp-allinone-http/0-install | test step completed 0-install logger.go:42: 14:47:55 | collector-otlp-allinone-http/1-smoke-test | starting test step 1-smoke-test logger.go:42: 14:47:55 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 14:47:56 | collector-otlp-allinone-http/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 14:48:02 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 14:48:03 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 14:48:03 | collector-otlp-allinone-http/1-smoke-test | job.batch/report-span created logger.go:42: 14:48:03 | collector-otlp-allinone-http/1-smoke-test | job.batch/check-span created logger.go:42: 14:48:14 | collector-otlp-allinone-http/1-smoke-test | test step completed 1-smoke-test logger.go:42: 14:48:14 | collector-otlp-allinone-http | collector-otlp-allinone-http events from ns kuttl-test-grateful-seal: logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:52 +0000 UTC Normal Pod my-jaeger-5f4fd48bc6-vnncv Binding Scheduled Successfully assigned kuttl-test-grateful-seal/my-jaeger-5f4fd48bc6-vnncv to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:52 +0000 UTC Normal ReplicaSet.apps my-jaeger-5f4fd48bc6 SuccessfulCreate Created pod: my-jaeger-5f4fd48bc6-vnncv replicaset-controller logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:52 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5f4fd48bc6 to 1 deployment-controller logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:53 +0000 UTC Warning Pod my-jaeger-5f4fd48bc6-vnncv FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-oauth-proxy-tls" : secret "my-jaeger-ui-oauth-proxy-tls" not found kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:53 +0000 UTC Warning Pod my-jaeger-5f4fd48bc6-vnncv FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:53 +0000 UTC Normal Pod my-jaeger-5f4fd48bc6-vnncv AddedInterface Add eth0 [10.128.2.54/23] from ovn-kubernetes logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:53 +0000 UTC Normal Pod my-jaeger-5f4fd48bc6-vnncv.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:54 +0000 UTC Normal Pod my-jaeger-5f4fd48bc6-vnncv.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:54 +0000 UTC Normal Pod my-jaeger-5f4fd48bc6-vnncv.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:54 +0000 UTC Normal Pod my-jaeger-5f4fd48bc6-vnncv.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:54 +0000 UTC Normal Pod my-jaeger-5f4fd48bc6-vnncv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:54 +0000 UTC Normal Pod my-jaeger-5f4fd48bc6-vnncv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:57 +0000 UTC Normal ReplicaSet.apps my-jaeger-5f4fd48bc6 SuccessfulDelete Deleted pod: my-jaeger-5f4fd48bc6-vnncv replicaset-controller logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:57 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-5f4fd48bc6 to 0 from 1 deployment-controller logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:58 +0000 UTC Normal Pod my-jaeger-5f4fd48bc6-vnncv.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:58 +0000 UTC Normal Pod my-jaeger-5f4fd48bc6-vnncv.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:59 +0000 UTC Normal Pod my-jaeger-6458668c9-rmflm Binding Scheduled Successfully assigned kuttl-test-grateful-seal/my-jaeger-6458668c9-rmflm to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:59 +0000 UTC Normal Pod my-jaeger-6458668c9-rmflm AddedInterface Add eth0 [10.128.2.55/23] from ovn-kubernetes logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:59 +0000 UTC Normal Pod my-jaeger-6458668c9-rmflm.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:59 +0000 UTC Normal Pod my-jaeger-6458668c9-rmflm.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:59 +0000 UTC Normal ReplicaSet.apps my-jaeger-6458668c9 SuccessfulCreate Created pod: my-jaeger-6458668c9-rmflm replicaset-controller logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:47:59 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6458668c9 to 1 deployment-controller logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:48:00 +0000 UTC Normal Pod my-jaeger-6458668c9-rmflm.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:48:00 +0000 UTC Normal Pod my-jaeger-6458668c9-rmflm.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:48:00 +0000 UTC Normal Pod my-jaeger-6458668c9-rmflm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:48:00 +0000 UTC Normal Pod my-jaeger-6458668c9-rmflm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:48:03 +0000 UTC Normal Pod check-span-jhgxh Binding Scheduled Successfully assigned kuttl-test-grateful-seal/check-span-jhgxh to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:48:03 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-jhgxh job-controller logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:48:03 +0000 UTC Normal Pod report-span-hjg4m Binding Scheduled Successfully assigned kuttl-test-grateful-seal/report-span-hjg4m to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:48:03 +0000 UTC Normal Pod report-span-hjg4m AddedInterface Add eth0 [10.129.2.106/23] from ovn-kubernetes logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:48:03 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-hjg4m job-controller logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:48:04 +0000 UTC Normal Pod check-span-jhgxh AddedInterface Add eth0 [10.131.0.71/23] from ovn-kubernetes logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:48:04 +0000 UTC Normal Pod check-span-jhgxh.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:48:04 +0000 UTC Normal Pod check-span-jhgxh.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:48:04 +0000 UTC Normal Pod check-span-jhgxh.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:48:04 +0000 UTC Normal Pod report-span-hjg4m.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:48:04 +0000 UTC Normal Pod report-span-hjg4m.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:48:04 +0000 UTC Normal Pod report-span-hjg4m.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 14:48:14 | collector-otlp-allinone-http | 2023-11-27 14:48:14 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 14:48:14 | collector-otlp-allinone-http | Deleting namespace: kuttl-test-grateful-seal === CONT kuttl/harness/collector-autoscale logger.go:42: 14:48:26 | collector-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:48:26 | collector-autoscale | Ignoring wait-for-hpa.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:48:26 | collector-autoscale | Creating namespace: kuttl-test-healthy-hedgehog logger.go:42: 14:48:26 | collector-autoscale/1-install | starting test step 1-install logger.go:42: 14:48:26 | collector-autoscale/1-install | Jaeger:kuttl-test-healthy-hedgehog/simple-prod created logger.go:42: 14:49:01 | collector-autoscale/1-install | test step completed 1-install logger.go:42: 14:49:01 | collector-autoscale/2-wait-for-hpa | starting test step 2-wait-for-hpa logger.go:42: 14:49:01 | collector-autoscale/2-wait-for-hpa | running command: [sh -c ./wait-for-hpa.sh] logger.go:42: 14:49:01 | collector-autoscale/2-wait-for-hpa | Some HPA metrics are not known yet logger.go:42: 14:49:02 | collector-autoscale/2-wait-for-hpa | test step completed 2-wait-for-hpa logger.go:42: 14:49:02 | collector-autoscale/3- | starting test step 3- logger.go:42: 14:49:02 | collector-autoscale/3- | test step completed 3- logger.go:42: 14:49:02 | collector-autoscale | collector-autoscale events from ns kuttl-test-healthy-hedgehog: logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:48:32 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltesthealthyhedgehogsimpleprod-1-767dfb699d SuccessfulCreate Created pod: elasticsearch-cdm-kuttltesthealthyhedgehogsimpleprod-1-767svpqs replicaset-controller logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:48:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthealthyhedgehogsimpleprod-1-767svpqs Binding Scheduled Successfully assigned kuttl-test-healthy-hedgehog/elasticsearch-cdm-kuttltesthealthyhedgehogsimpleprod-1-767svpqs to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:48:32 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltesthealthyhedgehogsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltesthealthyhedgehogsimpleprod-1-767dfb699d to 1 deployment-controller logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:48:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthealthyhedgehogsimpleprod-1-767svpqs AddedInterface Add eth0 [10.128.2.56/23] from ovn-kubernetes logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:48:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthealthyhedgehogsimpleprod-1-767svpqs.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:56ea62bfb0ca36e19a7b21aff3676e49511f05f72da5e76d6427fd8240f328a8" already present on machine kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:48:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthealthyhedgehogsimpleprod-1-767svpqs.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:48:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthealthyhedgehogsimpleprod-1-767svpqs.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:48:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthealthyhedgehogsimpleprod-1-767svpqs.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:79427bea6b5c37894b9782c3821d8b9074838e606daa4a743b2ae060856fa98a" already present on machine kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:48:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthealthyhedgehogsimpleprod-1-767svpqs.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:48:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthealthyhedgehogsimpleprod-1-767svpqs.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:48:43 +0000 UTC Warning Pod elasticsearch-cdm-kuttltesthealthyhedgehogsimpleprod-1-767svpqs.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:48:48 +0000 UTC Warning Pod elasticsearch-cdm-kuttltesthealthyhedgehogsimpleprod-1-767svpqs.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:48:59 +0000 UTC Normal Pod simple-prod-collector-67686746b8-2kp9d Binding Scheduled Successfully assigned kuttl-test-healthy-hedgehog/simple-prod-collector-67686746b8-2kp9d to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:48:59 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-67686746b8 SuccessfulCreate Created pod: simple-prod-collector-67686746b8-2kp9d replicaset-controller logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:48:59 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-67686746b8 to 1 deployment-controller logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:48:59 +0000 UTC Normal Pod simple-prod-query-6c4d6cfdd9-p4jks Binding Scheduled Successfully assigned kuttl-test-healthy-hedgehog/simple-prod-query-6c4d6cfdd9-p4jks to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:48:59 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6c4d6cfdd9 SuccessfulCreate Created pod: simple-prod-query-6c4d6cfdd9-p4jks replicaset-controller logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:48:59 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6c4d6cfdd9 to 1 deployment-controller logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:49:00 +0000 UTC Normal Pod simple-prod-collector-67686746b8-2kp9d AddedInterface Add eth0 [10.129.2.107/23] from ovn-kubernetes logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:49:00 +0000 UTC Normal Pod simple-prod-collector-67686746b8-2kp9d.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" already present on machine kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:49:00 +0000 UTC Normal Pod simple-prod-collector-67686746b8-2kp9d.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:49:00 +0000 UTC Normal Pod simple-prod-collector-67686746b8-2kp9d.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:49:00 +0000 UTC Normal Pod simple-prod-query-6c4d6cfdd9-p4jks AddedInterface Add eth0 [10.129.2.108/23] from ovn-kubernetes logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:49:00 +0000 UTC Normal Pod simple-prod-query-6c4d6cfdd9-p4jks.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:49:00 +0000 UTC Normal Pod simple-prod-query-6c4d6cfdd9-p4jks.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:49:00 +0000 UTC Normal Pod simple-prod-query-6c4d6cfdd9-p4jks.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:49:00 +0000 UTC Normal Pod simple-prod-query-6c4d6cfdd9-p4jks.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:49:00 +0000 UTC Normal Pod simple-prod-query-6c4d6cfdd9-p4jks.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:49:00 +0000 UTC Normal Pod simple-prod-query-6c4d6cfdd9-p4jks.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:49:00 +0000 UTC Normal Pod simple-prod-query-6c4d6cfdd9-p4jks.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:49:00 +0000 UTC Normal Pod simple-prod-query-6c4d6cfdd9-p4jks.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:49:02 | collector-autoscale | 2023-11-27 14:49:00 +0000 UTC Normal Pod simple-prod-query-6c4d6cfdd9-p4jks.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:49:02 | collector-autoscale | Deleting namespace: kuttl-test-healthy-hedgehog === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (321.36s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.65s) --- PASS: kuttl/harness/collector-otlp-production-grpc (75.78s) --- PASS: kuttl/harness/set-custom-img (47.73s) --- PASS: kuttl/harness/collector-otlp-production-http (67.48s) --- PASS: kuttl/harness/collector-otlp-allinone-grpc (45.25s) --- PASS: kuttl/harness/collector-otlp-allinone-http (36.90s) --- PASS: kuttl/harness/collector-autoscale (42.52s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name miscellaneous --report --output /logs/artifacts/miscellaneous.xml ./artifacts/kuttl-report.xml time="2023-11-27T14:49:09Z" level=debug msg="Setting a new name for the test suites" time="2023-11-27T14:49:09Z" level=debug msg="Removing 'artifacts' TestCase" time="2023-11-27T14:49:09Z" level=debug msg="normalizing test case names" time="2023-11-27T14:49:09Z" level=debug msg="miscellaneous/artifacts -> miscellaneous_artifacts" time="2023-11-27T14:49:09Z" level=debug msg="miscellaneous/collector-otlp-production-grpc -> miscellaneous_collector_otlp_production_grpc" time="2023-11-27T14:49:09Z" level=debug msg="miscellaneous/set-custom-img -> miscellaneous_set_custom_img" time="2023-11-27T14:49:09Z" level=debug msg="miscellaneous/collector-otlp-production-http -> miscellaneous_collector_otlp_production_http" time="2023-11-27T14:49:09Z" level=debug msg="miscellaneous/collector-otlp-allinone-grpc -> miscellaneous_collector_otlp_allinone_grpc" time="2023-11-27T14:49:09Z" level=debug msg="miscellaneous/collector-otlp-allinone-http -> miscellaneous_collector_otlp_allinone_http" time="2023-11-27T14:49:09Z" level=debug msg="miscellaneous/collector-autoscale -> miscellaneous_collector_autoscale" +----------------------------------------------+--------+ | NAME | RESULT | +----------------------------------------------+--------+ | miscellaneous_artifacts | passed | | miscellaneous_collector_otlp_production_grpc | passed | | miscellaneous_set_custom_img | passed | | miscellaneous_collector_otlp_production_http | passed | | miscellaneous_collector_otlp_allinone_grpc | passed | | miscellaneous_collector_otlp_allinone_http | passed | | miscellaneous_collector_autoscale | passed | +----------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh sidecar false true + '[' 3 -ne 3 ']' + test_suite_name=sidecar + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/sidecar.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-sidecar make[2]: Entering directory '/tmp/jaeger-tests' ./tests/e2e/sidecar/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-11-25-110147 True False 52m Cluster version is 4.15.0-0.nightly-2023-11-25-110147' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-11-25-110147 True False 52m Cluster version is 4.15.0-0.nightly-2023-11-25-110147' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z '' ']' ++ KAFKA_USE_CUSTOM_PODSET=false ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/sidecar/render.sh ++ export SUITE_DIR=./tests/e2e/sidecar ++ SUITE_DIR=./tests/e2e/sidecar ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/sidecar ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + jaeger_service_name=order + start_test sidecar-deployment + '[' 1 -ne 1 ']' + test_name=sidecar-deployment + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-deployment' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-deployment\e[0m' Rendering files for test sidecar-deployment + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build + '[' _build '!=' _build ']' + mkdir -p sidecar-deployment + cd sidecar-deployment + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-namespace + '[' 1 -ne 1 ']' + test_name=sidecar-namespace + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-namespace' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-namespace\e[0m' Rendering files for test sidecar-namespace + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-deployment + '[' sidecar-deployment '!=' _build ']' + cd .. + mkdir -p sidecar-namespace + cd sidecar-namespace + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-skip-webhook + '[' 1 -ne 1 ']' + test_name=sidecar-skip-webhook + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-skip-webhook' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-skip-webhook\e[0m' Rendering files for test sidecar-skip-webhook + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-namespace + '[' sidecar-namespace '!=' _build ']' + cd .. + mkdir -p sidecar-skip-webhook + cd sidecar-skip-webhook + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running sidecar E2E tests' Running sidecar E2E tests + cd tests/e2e/sidecar/_build + set +e + KUBECONFIG=/tmp/kubeconfig-4018005699 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/sidecar-deployment === PAUSE kuttl/harness/sidecar-deployment === RUN kuttl/harness/sidecar-namespace === PAUSE kuttl/harness/sidecar-namespace === RUN kuttl/harness/sidecar-skip-webhook === PAUSE kuttl/harness/sidecar-skip-webhook === CONT kuttl/harness/artifacts logger.go:42: 14:49:16 | artifacts | Creating namespace: kuttl-test-bold-locust logger.go:42: 14:49:16 | artifacts | artifacts events from ns kuttl-test-bold-locust: logger.go:42: 14:49:16 | artifacts | Deleting namespace: kuttl-test-bold-locust === CONT kuttl/harness/sidecar-namespace logger.go:42: 14:49:22 | sidecar-namespace | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:49:22 | sidecar-namespace | Creating namespace: kuttl-test-helpful-crow logger.go:42: 14:49:22 | sidecar-namespace/0-install | starting test step 0-install logger.go:42: 14:49:22 | sidecar-namespace/0-install | Jaeger:kuttl-test-helpful-crow/agent-as-sidecar created logger.go:42: 14:49:27 | sidecar-namespace/0-install | test step completed 0-install logger.go:42: 14:49:27 | sidecar-namespace/1-install | starting test step 1-install logger.go:42: 14:49:27 | sidecar-namespace/1-install | Deployment:kuttl-test-helpful-crow/vertx-create-span-sidecar created logger.go:42: 14:49:28 | sidecar-namespace/1-install | test step completed 1-install logger.go:42: 14:49:28 | sidecar-namespace/2-enable-injection | starting test step 2-enable-injection logger.go:42: 14:49:28 | sidecar-namespace/2-enable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="true"] logger.go:42: 14:49:28 | sidecar-namespace/2-enable-injection | namespace/kuttl-test-helpful-crow annotate logger.go:42: 14:49:30 | sidecar-namespace/2-enable-injection | test step completed 2-enable-injection logger.go:42: 14:49:30 | sidecar-namespace/3-find-service | starting test step 3-find-service logger.go:42: 14:49:30 | sidecar-namespace/3-find-service | Job:kuttl-test-helpful-crow/00-find-service created logger.go:42: 14:49:42 | sidecar-namespace/3-find-service | test step completed 3-find-service logger.go:42: 14:49:42 | sidecar-namespace/4-other-instance | starting test step 4-other-instance logger.go:42: 14:49:42 | sidecar-namespace/4-other-instance | Jaeger:kuttl-test-helpful-crow/agent-as-sidecar2 created logger.go:42: 14:49:50 | sidecar-namespace/4-other-instance | test step completed 4-other-instance logger.go:42: 14:49:50 | sidecar-namespace/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 14:49:51 | sidecar-namespace/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 14:49:51 | sidecar-namespace/6-find-service | starting test step 6-find-service logger.go:42: 14:49:52 | sidecar-namespace/6-find-service | Job:kuttl-test-helpful-crow/01-find-service created logger.go:42: 14:50:11 | sidecar-namespace/6-find-service | test step completed 6-find-service logger.go:42: 14:50:11 | sidecar-namespace/7-disable-injection | starting test step 7-disable-injection logger.go:42: 14:50:11 | sidecar-namespace/7-disable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="false"] logger.go:42: 14:50:11 | sidecar-namespace/7-disable-injection | namespace/kuttl-test-helpful-crow annotate logger.go:42: 14:50:16 | sidecar-namespace/7-disable-injection | test step completed 7-disable-injection logger.go:42: 14:50:16 | sidecar-namespace | sidecar-namespace events from ns kuttl-test-helpful-crow: logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:25 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-6f77f74d57 to 1 deployment-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:26 +0000 UTC Normal Pod agent-as-sidecar-6f77f74d57-7wk5v Binding Scheduled Successfully assigned kuttl-test-helpful-crow/agent-as-sidecar-6f77f74d57-7wk5v to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:26 +0000 UTC Normal Pod agent-as-sidecar-6f77f74d57-7wk5v AddedInterface Add eth0 [10.129.2.109/23] from ovn-kubernetes logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:26 +0000 UTC Normal Pod agent-as-sidecar-6f77f74d57-7wk5v.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:26 +0000 UTC Normal Pod agent-as-sidecar-6f77f74d57-7wk5v.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:26 +0000 UTC Normal Pod agent-as-sidecar-6f77f74d57-7wk5v.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:26 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-6f77f74d57 SuccessfulCreate Created pod: agent-as-sidecar-6f77f74d57-7wk5v replicaset-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:27 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-bpqn8 Binding Scheduled Successfully assigned kuttl-test-helpful-crow/vertx-create-span-sidecar-84d458b68c-bpqn8 to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:27 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-bpqn8 AddedInterface Add eth0 [10.128.2.57/23] from ovn-kubernetes logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:27 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-bpqn8 replicaset-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:27 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:28 +0000 UTC Normal Pod vertx-create-span-sidecar-5f4b557d95-tb4dc Binding Scheduled Successfully assigned kuttl-test-helpful-crow/vertx-create-span-sidecar-5f4b557d95-tb4dc to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:28 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-5f4b557d95 SuccessfulCreate Created pod: vertx-create-span-sidecar-5f4b557d95-tb4dc replicaset-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:28 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-bpqn8.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:28 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-bpqn8.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:28 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-bpqn8.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:28 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-5f4b557d95 to 1 deployment-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:29 +0000 UTC Normal Pod vertx-create-span-sidecar-5f4b557d95-tb4dc AddedInterface Add eth0 [10.129.2.110/23] from ovn-kubernetes logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:29 +0000 UTC Normal Pod vertx-create-span-sidecar-5f4b557d95-tb4dc.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:29 +0000 UTC Normal Pod vertx-create-span-sidecar-5f4b557d95-tb4dc.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:29 +0000 UTC Normal Pod vertx-create-span-sidecar-5f4b557d95-tb4dc.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:29 +0000 UTC Normal Pod vertx-create-span-sidecar-5f4b557d95-tb4dc.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:29 +0000 UTC Normal Pod vertx-create-span-sidecar-5f4b557d95-tb4dc.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:29 +0000 UTC Normal Pod vertx-create-span-sidecar-5f4b557d95-tb4dc.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:30 +0000 UTC Normal Pod 00-find-service-wptbz Binding Scheduled Successfully assigned kuttl-test-helpful-crow/00-find-service-wptbz to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:30 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-wptbz job-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:31 +0000 UTC Normal Pod 00-find-service-wptbz AddedInterface Add eth0 [10.131.0.72/23] from ovn-kubernetes logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:31 +0000 UTC Normal Pod 00-find-service-wptbz.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:31 +0000 UTC Normal Pod 00-find-service-wptbz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:31 +0000 UTC Normal Pod 00-find-service-wptbz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:36 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-bpqn8.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.57:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:36 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-bpqn8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.57:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:37 +0000 UTC Warning Pod vertx-create-span-sidecar-5f4b557d95-tb4dc.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.110:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:37 +0000 UTC Warning Pod vertx-create-span-sidecar-5f4b557d95-tb4dc.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.110:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:39 +0000 UTC Normal Pod vertx-create-span-sidecar-5f4b557d95-tb4dc.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:39 +0000 UTC Warning Pod vertx-create-span-sidecar-5f4b557d95-tb4dc.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.110:8080/": read tcp 10.129.2.2:39940->10.129.2.110:8080: read: connection reset by peer kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:39 +0000 UTC Warning Pod vertx-create-span-sidecar-5f4b557d95-tb4dc.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.110:8080/": dial tcp 10.129.2.110:8080: connect: connection refused kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:39 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-bpqn8.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:39 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-bpqn8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.57:8080/": read tcp 10.128.2.2:54804->10.128.2.57:8080: read: connection reset by peer kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:39 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-bpqn8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.57:8080/": dial tcp 10.128.2.57:8080: connect: connection refused kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:42 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:48 +0000 UTC Normal Pod agent-as-sidecar2-8fc964b9f-9zmgh Binding Scheduled Successfully assigned kuttl-test-helpful-crow/agent-as-sidecar2-8fc964b9f-9zmgh to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:48 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-8fc964b9f SuccessfulCreate Created pod: agent-as-sidecar2-8fc964b9f-9zmgh replicaset-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:48 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-8fc964b9f to 1 deployment-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:49 +0000 UTC Normal Pod agent-as-sidecar2-8fc964b9f-9zmgh AddedInterface Add eth0 [10.128.2.58/23] from ovn-kubernetes logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:49 +0000 UTC Normal Pod agent-as-sidecar2-8fc964b9f-9zmgh.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:49 +0000 UTC Normal Pod agent-as-sidecar2-8fc964b9f-9zmgh.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:49 +0000 UTC Normal Pod agent-as-sidecar2-8fc964b9f-9zmgh.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:50 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-bpqn8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.57:8080/": read tcp 10.128.2.2:46582->10.128.2.57:8080: read: connection reset by peer kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:51 +0000 UTC Normal Pod agent-as-sidecar-6f77f74d57-7wk5v.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:52 +0000 UTC Normal Pod 01-find-service-86bjb Binding Scheduled Successfully assigned kuttl-test-helpful-crow/01-find-service-86bjb to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:52 +0000 UTC Normal Pod 01-find-service-86bjb AddedInterface Add eth0 [10.131.0.73/23] from ovn-kubernetes logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:52 +0000 UTC Normal Pod 01-find-service-86bjb.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:52 +0000 UTC Normal Pod 01-find-service-86bjb.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:52 +0000 UTC Normal Pod 01-find-service-86bjb.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:52 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-86bjb job-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:53 +0000 UTC Normal Pod vertx-create-span-sidecar-7ff85b86d5-fjwbx Binding Scheduled Successfully assigned kuttl-test-helpful-crow/vertx-create-span-sidecar-7ff85b86d5-fjwbx to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:53 +0000 UTC Normal Pod vertx-create-span-sidecar-7ff85b86d5-fjwbx AddedInterface Add eth0 [10.128.2.59/23] from ovn-kubernetes logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:53 +0000 UTC Normal Pod vertx-create-span-sidecar-7ff85b86d5-fjwbx.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:53 +0000 UTC Normal Pod vertx-create-span-sidecar-7ff85b86d5-fjwbx.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:53 +0000 UTC Normal Pod vertx-create-span-sidecar-7ff85b86d5-fjwbx.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:53 +0000 UTC Normal Pod vertx-create-span-sidecar-7ff85b86d5-fjwbx.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:53 +0000 UTC Normal Pod vertx-create-span-sidecar-7ff85b86d5-fjwbx.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:53 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-7ff85b86d5 SuccessfulCreate Created pod: vertx-create-span-sidecar-7ff85b86d5-fjwbx replicaset-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:53 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-bpqn8 replicaset-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:53 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:53 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-7ff85b86d5 to 1 from 0 deployment-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:49:54 +0000 UTC Normal Pod vertx-create-span-sidecar-7ff85b86d5-fjwbx.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:50:01 +0000 UTC Warning Pod vertx-create-span-sidecar-7ff85b86d5-fjwbx.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.59:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:50:01 +0000 UTC Warning Pod vertx-create-span-sidecar-7ff85b86d5-fjwbx.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.59:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:50:03 +0000 UTC Normal Pod vertx-create-span-sidecar-7ff85b86d5-fjwbx.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:50:04 +0000 UTC Warning Pod vertx-create-span-sidecar-7ff85b86d5-fjwbx.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.59:8080/": read tcp 10.128.2.2:52038->10.128.2.59:8080: read: connection reset by peer kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:50:04 +0000 UTC Warning Pod vertx-create-span-sidecar-7ff85b86d5-fjwbx.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.59:8080/": dial tcp 10.128.2.59:8080: connect: connection refused kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:50:10 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:50:11 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-5f4b557d95 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-5f4b557d95-tb4dc replicaset-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:50:11 +0000 UTC Normal Pod vertx-create-span-sidecar-68c95f6996-v6n7j Binding Scheduled Successfully assigned kuttl-test-helpful-crow/vertx-create-span-sidecar-68c95f6996-v6n7j to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:50:11 +0000 UTC Normal Pod vertx-create-span-sidecar-68c95f6996-v6n7j AddedInterface Add eth0 [10.131.0.74/23] from ovn-kubernetes logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:50:11 +0000 UTC Normal Pod vertx-create-span-sidecar-68c95f6996-v6n7j.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:50:11 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-68c95f6996 SuccessfulCreate Created pod: vertx-create-span-sidecar-68c95f6996-v6n7j replicaset-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:50:11 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-5f4b557d95 to 0 from 1 deployment-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:50:11 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-68c95f6996 to 1 from 0 deployment-controller logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:50:15 +0000 UTC Normal Pod vertx-create-span-sidecar-68c95f6996-v6n7j.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.638s (3.638s including waiting) kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:50:15 +0000 UTC Normal Pod vertx-create-span-sidecar-68c95f6996-v6n7j.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 14:50:16 | sidecar-namespace | 2023-11-27 14:50:15 +0000 UTC Normal Pod vertx-create-span-sidecar-68c95f6996-v6n7j.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 14:50:16 | sidecar-namespace | Deleting namespace: kuttl-test-helpful-crow === CONT kuttl/harness/sidecar-skip-webhook logger.go:42: 14:50:22 | sidecar-skip-webhook | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:50:22 | sidecar-skip-webhook | Creating namespace: kuttl-test-thorough-mole logger.go:42: 14:50:22 | sidecar-skip-webhook/0-install | starting test step 0-install logger.go:42: 14:50:22 | sidecar-skip-webhook/0-install | Jaeger:kuttl-test-thorough-mole/agent-as-sidecar created logger.go:42: 14:50:27 | sidecar-skip-webhook/0-install | test step completed 0-install logger.go:42: 14:50:27 | sidecar-skip-webhook/1-install | starting test step 1-install logger.go:42: 14:50:27 | sidecar-skip-webhook/1-install | Deployment:kuttl-test-thorough-mole/vertx-create-span-sidecar created logger.go:42: 14:50:29 | sidecar-skip-webhook/1-install | test step completed 1-install logger.go:42: 14:50:29 | sidecar-skip-webhook/2-add-anotation-and-label | starting test step 2-add-anotation-and-label logger.go:42: 14:50:29 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name=jaeger-operator --namespace kuttl-test-thorough-mole] logger.go:42: 14:50:29 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar labeled logger.go:42: 14:50:29 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-thorough-mole] logger.go:42: 14:50:30 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar annotate logger.go:42: 14:50:30 | sidecar-skip-webhook/2-add-anotation-and-label | test step completed 2-add-anotation-and-label logger.go:42: 14:50:30 | sidecar-skip-webhook/3-remove-label | starting test step 3-remove-label logger.go:42: 14:50:30 | sidecar-skip-webhook/3-remove-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name- --namespace kuttl-test-thorough-mole] logger.go:42: 14:50:30 | sidecar-skip-webhook/3-remove-label | deployment.apps/vertx-create-span-sidecar unlabeled logger.go:42: 14:50:32 | sidecar-skip-webhook/3-remove-label | test step completed 3-remove-label logger.go:42: 14:50:32 | sidecar-skip-webhook | sidecar-skip-webhook events from ns kuttl-test-thorough-mole: logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:26 +0000 UTC Normal Pod agent-as-sidecar-7d9fcf5d85-76x6l Binding Scheduled Successfully assigned kuttl-test-thorough-mole/agent-as-sidecar-7d9fcf5d85-76x6l to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:26 +0000 UTC Normal Pod agent-as-sidecar-7d9fcf5d85-76x6l AddedInterface Add eth0 [10.128.2.60/23] from ovn-kubernetes logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:26 +0000 UTC Normal Pod agent-as-sidecar-7d9fcf5d85-76x6l.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:26 +0000 UTC Normal Pod agent-as-sidecar-7d9fcf5d85-76x6l.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:26 +0000 UTC Normal Pod agent-as-sidecar-7d9fcf5d85-76x6l.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:26 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-7d9fcf5d85 SuccessfulCreate Created pod: agent-as-sidecar-7d9fcf5d85-76x6l replicaset-controller logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:26 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-7d9fcf5d85 to 1 deployment-controller logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:27 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-nsl2v Binding Scheduled Successfully assigned kuttl-test-thorough-mole/vertx-create-span-sidecar-84d458b68c-nsl2v to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:27 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-nsl2v replicaset-controller logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:27 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:28 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-nsl2v AddedInterface Add eth0 [10.129.2.111/23] from ovn-kubernetes logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:28 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-nsl2v.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:28 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-nsl2v.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:28 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-nsl2v.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:30 +0000 UTC Normal Pod vertx-create-span-sidecar-67686674f7-ttz7w Binding Scheduled Successfully assigned kuttl-test-thorough-mole/vertx-create-span-sidecar-67686674f7-ttz7w to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:30 +0000 UTC Normal Pod vertx-create-span-sidecar-67686674f7-ttz7w AddedInterface Add eth0 [10.129.2.112/23] from ovn-kubernetes logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:30 +0000 UTC Normal Pod vertx-create-span-sidecar-67686674f7-ttz7w.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:30 +0000 UTC Normal Pod vertx-create-span-sidecar-67686674f7-ttz7w.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:30 +0000 UTC Normal Pod vertx-create-span-sidecar-67686674f7-ttz7w.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:30 +0000 UTC Normal Pod vertx-create-span-sidecar-67686674f7-ttz7w.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:30 +0000 UTC Normal Pod vertx-create-span-sidecar-67686674f7-ttz7w.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:30 +0000 UTC Normal Pod vertx-create-span-sidecar-67686674f7-ttz7w.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:30 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-67686674f7 SuccessfulCreate Created pod: vertx-create-span-sidecar-67686674f7-ttz7w replicaset-controller logger.go:42: 14:50:32 | sidecar-skip-webhook | 2023-11-27 14:50:30 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-67686674f7 to 1 deployment-controller logger.go:42: 14:50:32 | sidecar-skip-webhook | Deleting namespace: kuttl-test-thorough-mole === CONT kuttl/harness/sidecar-deployment logger.go:42: 14:50:38 | sidecar-deployment | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:50:38 | sidecar-deployment | Creating namespace: kuttl-test-square-fly logger.go:42: 14:50:38 | sidecar-deployment/0-install | starting test step 0-install logger.go:42: 14:50:38 | sidecar-deployment/0-install | Jaeger:kuttl-test-square-fly/agent-as-sidecar created logger.go:42: 14:50:44 | sidecar-deployment/0-install | test step completed 0-install logger.go:42: 14:50:44 | sidecar-deployment/1-install | starting test step 1-install logger.go:42: 14:50:44 | sidecar-deployment/1-install | Deployment:kuttl-test-square-fly/vertx-create-span-sidecar created logger.go:42: 14:50:46 | sidecar-deployment/1-install | test step completed 1-install logger.go:42: 14:50:46 | sidecar-deployment/2-enable-injection | starting test step 2-enable-injection logger.go:42: 14:50:46 | sidecar-deployment/2-enable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-square-fly] logger.go:42: 14:50:46 | sidecar-deployment/2-enable-injection | deployment.apps/vertx-create-span-sidecar annotate logger.go:42: 14:50:48 | sidecar-deployment/2-enable-injection | test step completed 2-enable-injection logger.go:42: 14:50:48 | sidecar-deployment/3-find-service | starting test step 3-find-service logger.go:42: 14:50:48 | sidecar-deployment/3-find-service | Job:kuttl-test-square-fly/00-find-service created logger.go:42: 14:51:00 | sidecar-deployment/3-find-service | test step completed 3-find-service logger.go:42: 14:51:00 | sidecar-deployment/4-other-instance | starting test step 4-other-instance logger.go:42: 14:51:00 | sidecar-deployment/4-other-instance | Jaeger:kuttl-test-square-fly/agent-as-sidecar2 created logger.go:42: 14:51:06 | sidecar-deployment/4-other-instance | test step completed 4-other-instance logger.go:42: 14:51:06 | sidecar-deployment/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 14:51:07 | sidecar-deployment/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 14:51:07 | sidecar-deployment/6-find-service | starting test step 6-find-service logger.go:42: 14:51:07 | sidecar-deployment/6-find-service | Job:kuttl-test-square-fly/01-find-service created logger.go:42: 14:51:19 | sidecar-deployment/6-find-service | test step completed 6-find-service logger.go:42: 14:51:19 | sidecar-deployment/7-disable-injection | starting test step 7-disable-injection logger.go:42: 14:51:19 | sidecar-deployment/7-disable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=false --namespace kuttl-test-square-fly] logger.go:42: 14:51:19 | sidecar-deployment/7-disable-injection | deployment.apps/vertx-create-span-sidecar annotate logger.go:42: 14:51:20 | sidecar-deployment/7-disable-injection | test step completed 7-disable-injection logger.go:42: 14:51:20 | sidecar-deployment | sidecar-deployment events from ns kuttl-test-square-fly: logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:41 +0000 UTC Normal Pod agent-as-sidecar-686bdcc9f4-5wmvj Binding Scheduled Successfully assigned kuttl-test-square-fly/agent-as-sidecar-686bdcc9f4-5wmvj to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:41 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-686bdcc9f4 SuccessfulCreate Created pod: agent-as-sidecar-686bdcc9f4-5wmvj replicaset-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:41 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-686bdcc9f4 to 1 deployment-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:42 +0000 UTC Normal Pod agent-as-sidecar-686bdcc9f4-5wmvj AddedInterface Add eth0 [10.128.2.61/23] from ovn-kubernetes logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:42 +0000 UTC Normal Pod agent-as-sidecar-686bdcc9f4-5wmvj.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:42 +0000 UTC Normal Pod agent-as-sidecar-686bdcc9f4-5wmvj.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:42 +0000 UTC Normal Pod agent-as-sidecar-686bdcc9f4-5wmvj.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:44 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-wjmzw Binding Scheduled Successfully assigned kuttl-test-square-fly/vertx-create-span-sidecar-84d458b68c-wjmzw to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:44 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-wjmzw AddedInterface Add eth0 [10.129.2.113/23] from ovn-kubernetes logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:44 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-wjmzw.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:44 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-wjmzw.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:44 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-wjmzw.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:44 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-wjmzw replicaset-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:44 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:46 +0000 UTC Normal Pod vertx-create-span-sidecar-d689d7776-sgvns Binding Scheduled Successfully assigned kuttl-test-square-fly/vertx-create-span-sidecar-d689d7776-sgvns to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:46 +0000 UTC Normal Pod vertx-create-span-sidecar-d689d7776-sgvns AddedInterface Add eth0 [10.129.2.114/23] from ovn-kubernetes logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:46 +0000 UTC Normal Pod vertx-create-span-sidecar-d689d7776-sgvns.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:46 +0000 UTC Normal Pod vertx-create-span-sidecar-d689d7776-sgvns.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:46 +0000 UTC Normal Pod vertx-create-span-sidecar-d689d7776-sgvns.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:46 +0000 UTC Normal Pod vertx-create-span-sidecar-d689d7776-sgvns.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:46 +0000 UTC Normal Pod vertx-create-span-sidecar-d689d7776-sgvns.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:46 +0000 UTC Normal Pod vertx-create-span-sidecar-d689d7776-sgvns.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:46 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-d689d7776 SuccessfulCreate Created pod: vertx-create-span-sidecar-d689d7776-sgvns replicaset-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:46 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-d689d7776 to 1 deployment-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:48 +0000 UTC Normal Pod 00-find-service-c4d9s Binding Scheduled Successfully assigned kuttl-test-square-fly/00-find-service-c4d9s to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:48 +0000 UTC Normal Pod 00-find-service-c4d9s AddedInterface Add eth0 [10.131.0.75/23] from ovn-kubernetes logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:48 +0000 UTC Normal Pod 00-find-service-c4d9s.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:48 +0000 UTC Normal Pod 00-find-service-c4d9s.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:48 +0000 UTC Normal Pod 00-find-service-c4d9s.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:48 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-c4d9s job-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:52 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-wjmzw.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.113:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:52 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-wjmzw.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.113:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:54 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-wjmzw.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:54 +0000 UTC Warning Pod vertx-create-span-sidecar-d689d7776-sgvns.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.114:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:54 +0000 UTC Warning Pod vertx-create-span-sidecar-d689d7776-sgvns.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.114:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:55 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-wjmzw.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.113:8080/": read tcp 10.129.2.2:47424->10.129.2.113:8080: read: connection reset by peer kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:55 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-wjmzw.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.113:8080/": dial tcp 10.129.2.113:8080: connect: connection refused kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:57 +0000 UTC Normal Pod vertx-create-span-sidecar-d689d7776-sgvns.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:57 +0000 UTC Warning Pod vertx-create-span-sidecar-d689d7776-sgvns.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.114:8080/": read tcp 10.129.2.2:47722->10.129.2.114:8080: read: connection reset by peer kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:57 +0000 UTC Warning Pod vertx-create-span-sidecar-d689d7776-sgvns.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.114:8080/": dial tcp 10.129.2.114:8080: connect: connection refused kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:50:59 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:04 +0000 UTC Normal Pod agent-as-sidecar2-77dc7cc78b-ptntj Binding Scheduled Successfully assigned kuttl-test-square-fly/agent-as-sidecar2-77dc7cc78b-ptntj to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:04 +0000 UTC Normal Pod agent-as-sidecar2-77dc7cc78b-ptntj AddedInterface Add eth0 [10.128.2.62/23] from ovn-kubernetes logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:04 +0000 UTC Normal Pod agent-as-sidecar2-77dc7cc78b-ptntj.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:04 +0000 UTC Normal Pod agent-as-sidecar2-77dc7cc78b-ptntj.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:04 +0000 UTC Normal Pod agent-as-sidecar2-77dc7cc78b-ptntj.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:04 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-77dc7cc78b SuccessfulCreate Created pod: agent-as-sidecar2-77dc7cc78b-ptntj replicaset-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:04 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-77dc7cc78b to 1 deployment-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:06 +0000 UTC Normal Pod agent-as-sidecar-686bdcc9f4-5wmvj.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:06 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-wjmzw.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.113:8080/": read tcp 10.129.2.2:46890->10.129.2.113:8080: read: connection reset by peer kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:07 +0000 UTC Normal Pod 01-find-service-8p5bf Binding Scheduled Successfully assigned kuttl-test-square-fly/01-find-service-8p5bf to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:07 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-8p5bf job-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:08 +0000 UTC Normal Pod 01-find-service-8p5bf AddedInterface Add eth0 [10.131.0.76/23] from ovn-kubernetes logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:08 +0000 UTC Normal Pod 01-find-service-8p5bf.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-1i9j6n46/pipeline@sha256:2a4ae348c30c5e46a5c821c7b46288c7c8b11b42125f77a20022f0207e64cb2e" already present on machine kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:08 +0000 UTC Normal Pod 01-find-service-8p5bf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:08 +0000 UTC Normal Pod 01-find-service-8p5bf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:08 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-wjmzw replicaset-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:08 +0000 UTC Normal Pod vertx-create-span-sidecar-c978f5cb6-6n89n Binding Scheduled Successfully assigned kuttl-test-square-fly/vertx-create-span-sidecar-c978f5cb6-6n89n to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:08 +0000 UTC Normal Pod vertx-create-span-sidecar-c978f5cb6-6n89n AddedInterface Add eth0 [10.128.2.63/23] from ovn-kubernetes logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:08 +0000 UTC Normal Pod vertx-create-span-sidecar-c978f5cb6-6n89n.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:08 +0000 UTC Normal Pod vertx-create-span-sidecar-c978f5cb6-6n89n.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:08 +0000 UTC Normal Pod vertx-create-span-sidecar-c978f5cb6-6n89n.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:08 +0000 UTC Normal Pod vertx-create-span-sidecar-c978f5cb6-6n89n.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:08 +0000 UTC Normal Pod vertx-create-span-sidecar-c978f5cb6-6n89n.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:08 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-c978f5cb6 SuccessfulCreate Created pod: vertx-create-span-sidecar-c978f5cb6-6n89n replicaset-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:08 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:08 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-c978f5cb6 to 1 from 0 deployment-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:09 +0000 UTC Normal Pod vertx-create-span-sidecar-c978f5cb6-6n89n.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:16 +0000 UTC Warning Pod vertx-create-span-sidecar-c978f5cb6-6n89n.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.63:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:16 +0000 UTC Warning Pod vertx-create-span-sidecar-c978f5cb6-6n89n.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.63:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:18 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:18 +0000 UTC Normal Pod vertx-create-span-sidecar-c978f5cb6-6n89n.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:19 +0000 UTC Normal Pod vertx-create-span-sidecar-85dbf5fcd7-2ctc8 Binding Scheduled Successfully assigned kuttl-test-square-fly/vertx-create-span-sidecar-85dbf5fcd7-2ctc8 to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:19 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-85dbf5fcd7 SuccessfulCreate Created pod: vertx-create-span-sidecar-85dbf5fcd7-2ctc8 replicaset-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:19 +0000 UTC Warning Pod vertx-create-span-sidecar-c978f5cb6-6n89n.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.63:8080/": read tcp 10.128.2.2:36946->10.128.2.63:8080: read: connection reset by peer kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:19 +0000 UTC Warning Pod vertx-create-span-sidecar-c978f5cb6-6n89n.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.63:8080/": dial tcp 10.128.2.63:8080: connect: connection refused kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:19 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-d689d7776 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-d689d7776-sgvns replicaset-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:19 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-d689d7776 to 0 from 1 deployment-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:19 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-85dbf5fcd7 to 1 from 0 deployment-controller logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:20 +0000 UTC Normal Pod vertx-create-span-sidecar-85dbf5fcd7-2ctc8 AddedInterface Add eth0 [10.131.0.77/23] from ovn-kubernetes logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:20 +0000 UTC Normal Pod vertx-create-span-sidecar-85dbf5fcd7-2ctc8.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:20 +0000 UTC Normal Pod vertx-create-span-sidecar-85dbf5fcd7-2ctc8.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 14:51:20 | sidecar-deployment | 2023-11-27 14:51:20 +0000 UTC Normal Pod vertx-create-span-sidecar-85dbf5fcd7-2ctc8.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 14:51:20 | sidecar-deployment | Deleting namespace: kuttl-test-square-fly === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (130.57s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.74s) --- PASS: kuttl/harness/sidecar-namespace (60.23s) --- PASS: kuttl/harness/sidecar-skip-webhook (15.41s) --- PASS: kuttl/harness/sidecar-deployment (49.14s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name sidecar --report --output /logs/artifacts/sidecar.xml ./artifacts/kuttl-report.xml time="2023-11-27T14:51:27Z" level=debug msg="Setting a new name for the test suites" time="2023-11-27T14:51:27Z" level=debug msg="Removing 'artifacts' TestCase" time="2023-11-27T14:51:27Z" level=debug msg="normalizing test case names" time="2023-11-27T14:51:27Z" level=debug msg="sidecar/artifacts -> sidecar_artifacts" time="2023-11-27T14:51:27Z" level=debug msg="sidecar/sidecar-namespace -> sidecar_sidecar_namespace" time="2023-11-27T14:51:27Z" level=debug msg="sidecar/sidecar-skip-webhook -> sidecar_sidecar_skip_webhook" time="2023-11-27T14:51:27Z" level=debug msg="sidecar/sidecar-deployment -> sidecar_sidecar_deployment" +------------------------------+--------+ | NAME | RESULT | +------------------------------+--------+ | sidecar_artifacts | passed | | sidecar_sidecar_namespace | passed | | sidecar_sidecar_skip_webhook | passed | | sidecar_sidecar_deployment | passed | +------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh streaming false true + '[' 3 -ne 3 ']' + test_suite_name=streaming + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/streaming.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-streaming make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=0.32.0 \ SKIP_KAFKA=false \ SKIP_ES_EXTERNAL=true \ ./tests/e2e/streaming/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-11-25-110147 True False 55m Cluster version is 4.15.0-0.nightly-2023-11-25-110147' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-11-25-110147 True False 55m Cluster version is 4.15.0-0.nightly-2023-11-25-110147' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 0.32.0 ']' ++ version_le 0.32.0 0.25.0 +++ echo 0.32.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 0.32.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/streaming/render.sh ++ export SUITE_DIR=./tests/e2e/streaming ++ SUITE_DIR=./tests/e2e/streaming ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/streaming ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + '[' false = true ']' + start_test streaming-simple + '[' 1 -ne 1 ']' + test_name=streaming-simple + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-simple' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-simple\e[0m' Rendering files for test streaming-simple + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + mkdir -p streaming-simple + cd streaming-simple + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./1-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/streaming-jaeger-assert.yaml.template -o ./04-assert.yaml + render_smoke_test simple-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=simple-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + export JAEGER_NAME=simple-streaming + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-tls + '[' 1 -ne 1 ']' + test_name=streaming-with-tls + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-tls' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-tls\e[0m' Rendering files for test streaming-with-tls + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-simple + '[' streaming-simple '!=' _build ']' + cd .. + mkdir -p streaming-with-tls + cd streaming-with-tls + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./1-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + render_smoke_test tls-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=tls-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + export JAEGER_NAME=tls-streaming + JAEGER_NAME=tls-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-autoprovisioning-autoscale + '[' 1 -ne 1 ']' + test_name=streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-autoprovisioning-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-autoprovisioning-autoscale\e[0m' Rendering files for test streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-tls + '[' streaming-with-tls '!=' _build ']' + cd .. + mkdir -p streaming-with-autoprovisioning-autoscale + cd streaming-with-autoprovisioning-autoscale + '[' true = true ']' + rm ./00-install.yaml ./00-assert.yaml + render_install_elasticsearch upstream 01 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=01 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./01-assert.yaml + jaeger_name=auto-provisioned + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="20Mi"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="500m"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.autoscale=true ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.minReplicas=1 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.maxReplicas=2 ./02-install.yaml + render_assert_kafka true auto-provisioned 03 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provisioned + test_step=03 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./03-assert.yaml ++ expr 03 + 1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./4-assert.yaml ++ expr 03 + 2 + CLUSTER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./05-assert.yaml + render_install_tracegen auto-provisioned 06 + '[' 2 -ne 2 ']' + jaeger=auto-provisioned + step=06 + replicas=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/tracegen.yaml -o ./06-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.replicas=1 ./06-install.yaml + sed -i s~simple-prod~auto-provisioned~gi ./06-install.yaml + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-tracegen.yaml.template -o ./06-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running streaming E2E tests' Running streaming E2E tests + cd tests/e2e/streaming/_build + set +e + KUBECONFIG=/tmp/kubeconfig-4018005699 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/streaming-simple === PAUSE kuttl/harness/streaming-simple === RUN kuttl/harness/streaming-with-autoprovisioning-autoscale === PAUSE kuttl/harness/streaming-with-autoprovisioning-autoscale === RUN kuttl/harness/streaming-with-tls === PAUSE kuttl/harness/streaming-with-tls === CONT kuttl/harness/artifacts logger.go:42: 14:51:40 | artifacts | Creating namespace: kuttl-test-precious-snail logger.go:42: 14:51:40 | artifacts | artifacts events from ns kuttl-test-precious-snail: logger.go:42: 14:51:40 | artifacts | Deleting namespace: kuttl-test-precious-snail === CONT kuttl/harness/streaming-with-autoprovisioning-autoscale logger.go:42: 14:51:46 | streaming-with-autoprovisioning-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:51:46 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:51:46 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:51:46 | streaming-with-autoprovisioning-autoscale | Creating namespace: kuttl-test-first-unicorn logger.go:42: 14:51:46 | streaming-with-autoprovisioning-autoscale/1-install | starting test step 1-install logger.go:42: 14:51:46 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc create sa deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 14:51:46 | streaming-with-autoprovisioning-autoscale/1-install | serviceaccount/deploy-elasticsearch created logger.go:42: 14:51:46 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc adm policy add-scc-to-user privileged -z deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 14:51:46 | streaming-with-autoprovisioning-autoscale/1-install | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:privileged added: "deploy-elasticsearch" logger.go:42: 14:51:46 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 6] logger.go:42: 14:51:52 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_0.yml -n $NAMESPACE] logger.go:42: 14:51:52 | streaming-with-autoprovisioning-autoscale/1-install | statefulset.apps/elasticsearch created logger.go:42: 14:51:52 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 3] logger.go:42: 14:51:55 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_1.yml -n $NAMESPACE] logger.go:42: 14:51:56 | streaming-with-autoprovisioning-autoscale/1-install | service/elasticsearch created logger.go:42: 14:52:13 | streaming-with-autoprovisioning-autoscale/1-install | test step completed 1-install logger.go:42: 14:52:13 | streaming-with-autoprovisioning-autoscale/2-install | starting test step 2-install logger.go:42: 14:52:13 | streaming-with-autoprovisioning-autoscale/2-install | Jaeger:kuttl-test-first-unicorn/auto-provisioned created logger.go:42: 14:52:13 | streaming-with-autoprovisioning-autoscale/2-install | test step completed 2-install logger.go:42: 14:52:13 | streaming-with-autoprovisioning-autoscale/3- | starting test step 3- logger.go:42: 14:52:49 | streaming-with-autoprovisioning-autoscale/3- | test step completed 3- logger.go:42: 14:52:49 | streaming-with-autoprovisioning-autoscale/4- | starting test step 4- logger.go:42: 14:53:19 | streaming-with-autoprovisioning-autoscale/4- | test step completed 4- logger.go:42: 14:53:19 | streaming-with-autoprovisioning-autoscale/5- | starting test step 5- logger.go:42: 14:53:41 | streaming-with-autoprovisioning-autoscale/5- | test step completed 5- logger.go:42: 14:53:41 | streaming-with-autoprovisioning-autoscale/6-install | starting test step 6-install logger.go:42: 14:53:41 | streaming-with-autoprovisioning-autoscale/6-install | Deployment:kuttl-test-first-unicorn/tracegen created logger.go:42: 14:53:46 | streaming-with-autoprovisioning-autoscale/6-install | test step completed 6-install logger.go:42: 14:53:46 | streaming-with-autoprovisioning-autoscale/7- | starting test step 7- logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale/7- | test step completed 7- logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | streaming-with-autoprovisioning-autoscale events from ns kuttl-test-first-unicorn: logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:51:52 +0000 UTC Normal Pod elasticsearch-0 Binding Scheduled Successfully assigned kuttl-test-first-unicorn/elasticsearch-0 to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:51:52 +0000 UTC Normal StatefulSet.apps elasticsearch SuccessfulCreate create Pod elasticsearch-0 in StatefulSet elasticsearch successful statefulset-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:51:53 +0000 UTC Normal Pod elasticsearch-0 AddedInterface Add eth0 [10.128.2.64/23] from ovn-kubernetes logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:51:53 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulling Pulling image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:01 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulled Successfully pulled image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" in 8.63s (8.63s including waiting) kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:01 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:01 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:08 +0000 UTC Warning Pod elasticsearch-0.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Get "http://10.128.2.64:9200/": dial tcp 10.128.2.64:9200: connect: connection refused kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:17 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-zookeeper NoPods No matching pods found controllermanager logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:17 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:17 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:17 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-first-unicorn/data-auto-provisioned-zookeeper-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-684659d4f7-2wwfg_180f48e7-ecaa-4b6b-8bcf-6bd94c7ad5a7 logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:20 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ProvisioningSucceeded Successfully provisioned volume pvc-115be430-dcdf-450b-8b43-b49d0d59a521 ebs.csi.aws.com_aws-ebs-csi-driver-controller-684659d4f7-2wwfg_180f48e7-ecaa-4b6b-8bcf-6bd94c7ad5a7 logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:21 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 Binding Scheduled Successfully assigned kuttl-test-first-unicorn/auto-provisioned-zookeeper-0 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:23 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-115be430-dcdf-450b-8b43-b49d0d59a521" attachdetach-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:28 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 AddedInterface Add eth0 [10.129.2.115/23] from ovn-kubernetes logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:28 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulled Container image "registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:0bcdd55f01638f650ed69ebdf5f8a7291e103805b8cbb34013ced88e46e0678c" already present on machine kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:28 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Created Created container zookeeper kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:28 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Started Started container zookeeper kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:49 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:50 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-kafka NoPods No matching pods found controllermanager logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:50 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:50 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-first-unicorn/data-0-auto-provisioned-kafka-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-684659d4f7-2wwfg_180f48e7-ecaa-4b6b-8bcf-6bd94c7ad5a7 logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:53 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ProvisioningSucceeded Successfully provisioned volume pvc-02fd5013-069f-4408-8de0-4332d4ba2bf8 ebs.csi.aws.com_aws-ebs-csi-driver-controller-684659d4f7-2wwfg_180f48e7-ecaa-4b6b-8bcf-6bd94c7ad5a7 logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:54 +0000 UTC Normal Pod auto-provisioned-kafka-0 Binding Scheduled Successfully assigned kuttl-test-first-unicorn/auto-provisioned-kafka-0 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:56 +0000 UTC Normal Pod auto-provisioned-kafka-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-02fd5013-069f-4408-8de0-4332d4ba2bf8" attachdetach-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:59 +0000 UTC Normal Pod auto-provisioned-kafka-0 AddedInterface Add eth0 [10.129.2.116/23] from ovn-kubernetes logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:59 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulled Container image "registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:0bcdd55f01638f650ed69ebdf5f8a7291e103805b8cbb34013ced88e46e0678c" already present on machine kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:59 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Created Created container kafka kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:52:59 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Started Started container kafka kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:19 +0000 UTC Normal Pod auto-provisioned-entity-operator-66d4dc565b-6d5pc Binding Scheduled Successfully assigned kuttl-test-first-unicorn/auto-provisioned-entity-operator-66d4dc565b-6d5pc to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:19 +0000 UTC Normal ReplicaSet.apps auto-provisioned-entity-operator-66d4dc565b SuccessfulCreate Created pod: auto-provisioned-entity-operator-66d4dc565b-6d5pc replicaset-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:19 +0000 UTC Normal Deployment.apps auto-provisioned-entity-operator ScalingReplicaSet Scaled up replica set auto-provisioned-entity-operator-66d4dc565b to 1 deployment-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:20 +0000 UTC Normal Pod auto-provisioned-entity-operator-66d4dc565b-6d5pc AddedInterface Add eth0 [10.129.2.117/23] from ovn-kubernetes logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:20 +0000 UTC Normal Pod auto-provisioned-entity-operator-66d4dc565b-6d5pc.spec.containers{topic-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:464b04e622e0b3472e8a1e1ce8a2efd32cf27fc2056d3d589bfe6b5f9ac0bf4e" already present on machine kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:20 +0000 UTC Normal Pod auto-provisioned-entity-operator-66d4dc565b-6d5pc.spec.containers{topic-operator} Created Created container topic-operator kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:20 +0000 UTC Normal Pod auto-provisioned-entity-operator-66d4dc565b-6d5pc.spec.containers{topic-operator} Started Started container topic-operator kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:20 +0000 UTC Normal Pod auto-provisioned-entity-operator-66d4dc565b-6d5pc.spec.containers{user-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:464b04e622e0b3472e8a1e1ce8a2efd32cf27fc2056d3d589bfe6b5f9ac0bf4e" already present on machine kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:20 +0000 UTC Normal Pod auto-provisioned-entity-operator-66d4dc565b-6d5pc.spec.containers{user-operator} Created Created container user-operator kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:20 +0000 UTC Normal Pod auto-provisioned-entity-operator-66d4dc565b-6d5pc.spec.containers{user-operator} Started Started container user-operator kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:20 +0000 UTC Normal Pod auto-provisioned-entity-operator-66d4dc565b-6d5pc.spec.containers{tls-sidecar} Pulled Container image "registry.redhat.io/amq-streams/kafka-35-rhel8@sha256:0bcdd55f01638f650ed69ebdf5f8a7291e103805b8cbb34013ced88e46e0678c" already present on machine kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:20 +0000 UTC Normal Pod auto-provisioned-entity-operator-66d4dc565b-6d5pc.spec.containers{tls-sidecar} Created Created container tls-sidecar kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:20 +0000 UTC Normal Pod auto-provisioned-entity-operator-66d4dc565b-6d5pc.spec.containers{tls-sidecar} Started Started container tls-sidecar kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:41 +0000 UTC Normal Pod tracegen-5997856768-29fx5 Binding Scheduled Successfully assigned kuttl-test-first-unicorn/tracegen-5997856768-29fx5 to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:41 +0000 UTC Warning Pod tracegen-5997856768-29fx5 FailedMount MountVolume.SetUp failed for volume "auto-provisioned-service-ca" : configmap "auto-provisioned-service-ca" not found kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:41 +0000 UTC Warning Pod tracegen-5997856768-29fx5 FailedMount MountVolume.SetUp failed for volume "auto-provisioned-trusted-ca" : configmap "auto-provisioned-trusted-ca" not found kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:41 +0000 UTC Normal ReplicaSet.apps tracegen-5997856768 SuccessfulCreate Created pod: tracegen-5997856768-29fx5 replicaset-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:41 +0000 UTC Normal Deployment.apps tracegen ScalingReplicaSet Scaled up replica set tracegen-5997856768 to 1 deployment-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:42 +0000 UTC Normal Pod auto-provisioned-collector-5dcdc9fc89-mg7qb Binding Scheduled Successfully assigned kuttl-test-first-unicorn/auto-provisioned-collector-5dcdc9fc89-mg7qb to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:42 +0000 UTC Warning Pod auto-provisioned-collector-5dcdc9fc89-mg7qb FailedMount MountVolume.SetUp failed for volume "auto-provisioned-collector-tls-config-volume" : secret "auto-provisioned-collector-headless-tls" not found kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:42 +0000 UTC Normal ReplicaSet.apps auto-provisioned-collector-5dcdc9fc89 SuccessfulCreate Created pod: auto-provisioned-collector-5dcdc9fc89-mg7qb replicaset-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:42 +0000 UTC Normal Deployment.apps auto-provisioned-collector ScalingReplicaSet Scaled up replica set auto-provisioned-collector-5dcdc9fc89 to 1 deployment-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:42 +0000 UTC Normal Pod auto-provisioned-ingester-57b5fd467d-ds4vl Binding Scheduled Successfully assigned kuttl-test-first-unicorn/auto-provisioned-ingester-57b5fd467d-ds4vl to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:42 +0000 UTC Normal ReplicaSet.apps auto-provisioned-ingester-57b5fd467d SuccessfulCreate Created pod: auto-provisioned-ingester-57b5fd467d-ds4vl replicaset-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:42 +0000 UTC Normal Deployment.apps auto-provisioned-ingester ScalingReplicaSet Scaled up replica set auto-provisioned-ingester-57b5fd467d to 1 deployment-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:42 +0000 UTC Normal Pod auto-provisioned-query-7b4f7f966c-266bs Binding Scheduled Successfully assigned kuttl-test-first-unicorn/auto-provisioned-query-7b4f7f966c-266bs to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:42 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-7b4f7f966c SuccessfulCreate Created pod: auto-provisioned-query-7b4f7f966c-266bs replicaset-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:42 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled up replica set auto-provisioned-query-7b4f7f966c to 1 deployment-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:43 +0000 UTC Normal Pod auto-provisioned-collector-5dcdc9fc89-mg7qb AddedInterface Add eth0 [10.131.0.78/23] from ovn-kubernetes logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:43 +0000 UTC Normal Pod auto-provisioned-collector-5dcdc9fc89-mg7qb.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" already present on machine kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:43 +0000 UTC Normal Pod auto-provisioned-collector-5dcdc9fc89-mg7qb.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:43 +0000 UTC Normal Pod auto-provisioned-collector-5dcdc9fc89-mg7qb.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:43 +0000 UTC Normal Pod auto-provisioned-ingester-57b5fd467d-ds4vl AddedInterface Add eth0 [10.128.2.66/23] from ovn-kubernetes logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:43 +0000 UTC Normal Pod auto-provisioned-ingester-57b5fd467d-ds4vl.spec.containers{jaeger-ingester} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:109a3676bfaccda0021c57bbe82bceed5140faaedb8bad4d7d2be7cd660de039" kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:43 +0000 UTC Normal Pod auto-provisioned-query-7b4f7f966c-266bs AddedInterface Add eth0 [10.131.0.79/23] from ovn-kubernetes logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:43 +0000 UTC Normal Pod auto-provisioned-query-7b4f7f966c-266bs.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:43 +0000 UTC Normal Pod auto-provisioned-query-7b4f7f966c-266bs.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:43 +0000 UTC Normal Pod auto-provisioned-query-7b4f7f966c-266bs.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:43 +0000 UTC Normal Pod auto-provisioned-query-7b4f7f966c-266bs.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:43 +0000 UTC Normal Pod auto-provisioned-query-7b4f7f966c-266bs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:43 +0000 UTC Normal Pod auto-provisioned-query-7b4f7f966c-266bs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:43 +0000 UTC Normal Pod auto-provisioned-query-7b4f7f966c-266bs.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:43 +0000 UTC Normal Pod auto-provisioned-query-7b4f7f966c-266bs.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:43 +0000 UTC Normal Pod auto-provisioned-query-7b4f7f966c-266bs.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:43 +0000 UTC Normal Pod tracegen-5997856768-29fx5 AddedInterface Add eth0 [10.128.2.65/23] from ovn-kubernetes logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:43 +0000 UTC Normal Pod tracegen-5997856768-29fx5.spec.containers{tracegen} Pulling Pulling image "jaegertracing/jaeger-tracegen:1.51.0" kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:44 +0000 UTC Normal Pod tracegen-5997856768-29fx5.spec.containers{tracegen} Pulled Successfully pulled image "jaegertracing/jaeger-tracegen:1.51.0" in 1.172s (1.172s including waiting) kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:44 +0000 UTC Normal Pod tracegen-5997856768-29fx5.spec.containers{tracegen} Created Created container tracegen kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:44 +0000 UTC Normal Pod tracegen-5997856768-29fx5.spec.containers{tracegen} Started Started container tracegen kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:44 +0000 UTC Normal Pod tracegen-5997856768-29fx5.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:44 +0000 UTC Normal Pod tracegen-5997856768-29fx5.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:44 +0000 UTC Normal Pod tracegen-5997856768-29fx5.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:46 +0000 UTC Normal Pod auto-provisioned-ingester-57b5fd467d-ds4vl.spec.containers{jaeger-ingester} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:109a3676bfaccda0021c57bbe82bceed5140faaedb8bad4d7d2be7cd660de039" in 2.817s (2.817s including waiting) kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:46 +0000 UTC Normal Pod auto-provisioned-ingester-57b5fd467d-ds4vl.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:46 +0000 UTC Normal Pod auto-provisioned-ingester-57b5fd467d-ds4vl.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:57 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:57 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:57 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:57 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-ingester FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:57 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-ingester FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:53:57 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-ingester FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:54:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod auto-provisioned-collector-5dcdc9fc89-mg7qb horizontal-pod-autoscaler logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:54:43 +0000 UTC Normal Pod auto-provisioned-ingester-57b5fd467d-w4lsf Binding Scheduled Successfully assigned kuttl-test-first-unicorn/auto-provisioned-ingester-57b5fd467d-w4lsf to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:54:43 +0000 UTC Normal Pod auto-provisioned-ingester-57b5fd467d-w4lsf AddedInterface Add eth0 [10.131.0.80/23] from ovn-kubernetes logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:54:43 +0000 UTC Normal Pod auto-provisioned-ingester-57b5fd467d-w4lsf.spec.containers{jaeger-ingester} Pulled Container image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:109a3676bfaccda0021c57bbe82bceed5140faaedb8bad4d7d2be7cd660de039" already present on machine kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:54:43 +0000 UTC Normal Pod auto-provisioned-ingester-57b5fd467d-w4lsf.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:54:43 +0000 UTC Normal Pod auto-provisioned-ingester-57b5fd467d-w4lsf.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:54:43 +0000 UTC Normal ReplicaSet.apps auto-provisioned-ingester-57b5fd467d SuccessfulCreate Created pod: auto-provisioned-ingester-57b5fd467d-w4lsf replicaset-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:54:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-ingester FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:54:43 +0000 UTC Normal HorizontalPodAutoscaler.autoscaling auto-provisioned-ingester SuccessfulRescale New size: 2; reason: memory resource utilization (percentage of request) above target horizontal-pod-autoscaler logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | 2023-11-27 14:54:43 +0000 UTC Normal Deployment.apps auto-provisioned-ingester ScalingReplicaSet Scaled up replica set auto-provisioned-ingester-57b5fd467d to 2 from 1 deployment-controller logger.go:42: 14:54:44 | streaming-with-autoprovisioning-autoscale | Deleting namespace: kuttl-test-first-unicorn === CONT kuttl/harness/streaming-with-tls logger.go:42: 14:55:02 | streaming-with-tls | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:55:02 | streaming-with-tls | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:55:02 | streaming-with-tls | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 14:55:02 | streaming-with-tls | Creating namespace: kuttl-test-super-mosquito logger.go:42: 14:55:02 | streaming-with-tls/0-install | starting test step 0-install logger.go:42: 14:55:02 | streaming-with-tls/0-install | running command: [sh -c cd /tmp/jaeger-tests && make undeploy-kafka KAFKA_NAMESPACE=$NAMESPACE KAFKA_OLM=true] logger.go:42: 14:55:02 | streaming-with-tls/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 14:55:02 | streaming-with-tls/0-install | >>>> Skiping kafka-operator undeploy logger.go:42: 14:55:02 | streaming-with-tls/0-install | kubectl delete --namespace kuttl-test-super-mosquito -f tests/_build/kafka-example.yaml 2>&1 || true logger.go:42: 14:55:02 | streaming-with-tls/0-install | error: the path "tests/_build/kafka-example.yaml" does not exist logger.go:42: 14:55:02 | streaming-with-tls/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 14:55:02 | streaming-with-tls/0-install | running command: [sh -c cd /tmp/jaeger-tests && make kafka KAFKA_NAMESPACE=$NAMESPACE KAFKA_OLM=true KAFKA_VERSION=0.32.0] logger.go:42: 14:55:02 | streaming-with-tls/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 14:55:02 | streaming-with-tls/0-install | >>>> Creating namespace kuttl-test-super-mosquito logger.go:42: 14:55:02 | streaming-with-tls/0-install | kubectl create namespace kuttl-test-super-mosquito 2>&1 | grep -v "already exists" || true logger.go:42: 14:55:02 | streaming-with-tls/0-install | >>>> Skipping kafka-operator deployment, assuming it has been installed via OperatorHub logger.go:42: 14:55:02 | streaming-with-tls/0-install | >>>> Creating namespace kuttl-test-super-mosquito logger.go:42: 14:55:02 | streaming-with-tls/0-install | mkdir -p tests/_build/ logger.go:42: 14:55:02 | streaming-with-tls/0-install | kubectl create namespace kuttl-test-super-mosquito 2>&1 | grep -v "already exists" || true logger.go:42: 14:55:02 | streaming-with-tls/0-install | curl --fail --location "https://raw.githubusercontent.com/strimzi/strimzi-kafka-operator/0.32.0/examples/kafka/kafka-persistent-single.yaml" --output tests/_build/kafka-example.yaml --create-dirs logger.go:42: 14:55:02 | streaming-with-tls/0-install | % Total % Received % Xferd Average Speed Time Time Time Current logger.go:42: 14:55:02 | streaming-with-tls/0-install | Dload Upload Total Spent Left Speed logger.go:42: 14:55:02 | streaming-with-tls/0-install | 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 865 100 865 0 0 6462 0 --:--:-- --:--:-- --:--:-- 6503 logger.go:42: 14:55:02 | streaming-with-tls/0-install | "sed" -i 's/size: 100Gi/size: 10Gi/g' tests/_build/kafka-example.yaml logger.go:42: 14:55:02 | streaming-with-tls/0-install | kubectl -n kuttl-test-super-mosquito apply --dry-run=client -f tests/_build/kafka-example.yaml logger.go:42: 14:55:02 | streaming-with-tls/0-install | kafka.kafka.strimzi.io/my-cluster created (dry run) logger.go:42: 14:55:02 | streaming-with-tls/0-install | kubectl -n kuttl-test-super-mosquito apply -f tests/_build/kafka-example.yaml 2>&1 | grep -v "already exists" || true logger.go:42: 14:55:03 | streaming-with-tls/0-install | kafka.kafka.strimzi.io/my-cluster created logger.go:42: 14:55:03 | streaming-with-tls/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 15:05:03 | streaming-with-tls/0-install | test step failed 0-install case.go:364: failed in step 0-install case.go:366: strimzipodsets.core.strimzi.io "my-cluster-zookeeper" not found logger.go:42: 15:05:03 | streaming-with-tls | streaming-with-tls events from ns kuttl-test-super-mosquito: logger.go:42: 15:05:03 | streaming-with-tls | Deleting namespace: kuttl-test-super-mosquito === CONT kuttl/harness/streaming-simple logger.go:42: 15:05:09 | streaming-simple | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 15:05:09 | streaming-simple | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 15:05:09 | streaming-simple | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 15:05:09 | streaming-simple | Creating namespace: kuttl-test-apparent-boar logger.go:42: 15:05:09 | streaming-simple/0-install | starting test step 0-install logger.go:42: 15:05:09 | streaming-simple/0-install | running command: [sh -c cd /tmp/jaeger-tests && make undeploy-kafka KAFKA_NAMESPACE=$NAMESPACE KAFKA_OLM=true] logger.go:42: 15:05:09 | streaming-simple/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 15:05:09 | streaming-simple/0-install | >>>> Skiping kafka-operator undeploy logger.go:42: 15:05:09 | streaming-simple/0-install | kubectl delete --namespace kuttl-test-apparent-boar -f tests/_build/kafka-example.yaml 2>&1 || true logger.go:42: 15:05:09 | streaming-simple/0-install | Error from server (NotFound): error when deleting "tests/_build/kafka-example.yaml": kafkas.kafka.strimzi.io "my-cluster" not found logger.go:42: 15:05:09 | streaming-simple/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 15:05:09 | streaming-simple/0-install | running command: [sh -c cd /tmp/jaeger-tests && make kafka KAFKA_NAMESPACE=$NAMESPACE KAFKA_OLM=true KAFKA_VERSION=0.32.0] logger.go:42: 15:05:09 | streaming-simple/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 15:05:09 | streaming-simple/0-install | >>>> Creating namespace kuttl-test-apparent-boar logger.go:42: 15:05:09 | streaming-simple/0-install | kubectl create namespace kuttl-test-apparent-boar 2>&1 | grep -v "already exists" || true logger.go:42: 15:05:09 | streaming-simple/0-install | >>>> Skipping kafka-operator deployment, assuming it has been installed via OperatorHub logger.go:42: 15:05:09 | streaming-simple/0-install | >>>> Creating namespace kuttl-test-apparent-boar logger.go:42: 15:05:09 | streaming-simple/0-install | mkdir -p tests/_build/ logger.go:42: 15:05:09 | streaming-simple/0-install | kubectl create namespace kuttl-test-apparent-boar 2>&1 | grep -v "already exists" || true logger.go:42: 15:05:09 | streaming-simple/0-install | curl --fail --location "https://raw.githubusercontent.com/strimzi/strimzi-kafka-operator/0.32.0/examples/kafka/kafka-persistent-single.yaml" --output tests/_build/kafka-example.yaml --create-dirs logger.go:42: 15:05:09 | streaming-simple/0-install | % Total % Received % Xferd Average Speed Time Time Time Current logger.go:42: 15:05:09 | streaming-simple/0-install | Dload Upload Total Spent Left Speed logger.go:42: 15:05:09 | streaming-simple/0-install | 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 865 100 865 0 0 6546 0 --:--:-- --:--:-- --:--:-- 6553 logger.go:42: 15:05:09 | streaming-simple/0-install | "sed" -i 's/size: 100Gi/size: 10Gi/g' tests/_build/kafka-example.yaml logger.go:42: 15:05:09 | streaming-simple/0-install | kubectl -n kuttl-test-apparent-boar apply --dry-run=client -f tests/_build/kafka-example.yaml logger.go:42: 15:05:09 | streaming-simple/0-install | kafka.kafka.strimzi.io/my-cluster created (dry run) logger.go:42: 15:05:09 | streaming-simple/0-install | kubectl -n kuttl-test-apparent-boar apply -f tests/_build/kafka-example.yaml 2>&1 | grep -v "already exists" || true logger.go:42: 15:05:10 | streaming-simple/0-install | kafka.kafka.strimzi.io/my-cluster created logger.go:42: 15:05:10 | streaming-simple/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 15:15:11 | streaming-simple/0-install | test step failed 0-install case.go:364: failed in step 0-install case.go:366: strimzipodsets.core.strimzi.io "my-cluster-zookeeper" not found logger.go:42: 15:15:11 | streaming-simple | streaming-simple events from ns kuttl-test-apparent-boar: logger.go:42: 15:15:11 | streaming-simple | Deleting namespace: kuttl-test-apparent-boar === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (1416.53s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.66s) --- PASS: kuttl/harness/streaming-with-autoprovisioning-autoscale (196.15s) --- FAIL: kuttl/harness/streaming-with-tls (606.81s) --- FAIL: kuttl/harness/streaming-simple (607.87s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name streaming --report --output /logs/artifacts/streaming.xml ./artifacts/kuttl-report.xml time="2023-11-27T15:15:17Z" level=debug msg="Setting a new name for the test suites" time="2023-11-27T15:15:17Z" level=debug msg="Removing 'artifacts' TestCase" time="2023-11-27T15:15:17Z" level=debug msg="normalizing test case names" time="2023-11-27T15:15:17Z" level=debug msg="streaming/artifacts -> streaming_artifacts" time="2023-11-27T15:15:17Z" level=debug msg="streaming/streaming-with-autoprovisioning-autoscale -> streaming_streaming_with_autoprovisioning_autoscale" time="2023-11-27T15:15:17Z" level=debug msg="streaming/streaming-with-tls -> streaming_streaming_with_tls" time="2023-11-27T15:15:17Z" level=debug msg="streaming/streaming-simple -> streaming_streaming_simple" +-----------------------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------------------+--------+ | streaming_artifacts | passed | | streaming_streaming_with_autoprovisioning_autoscale | passed | | streaming_streaming_with_tls | failed | | streaming_streaming_simple | failed | +-----------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh ui false true + '[' 3 -ne 3 ']' + test_suite_name=ui + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/ui.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-ui make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true ./tests/e2e/ui/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-11-25-110147 True False 79m Cluster version is 4.15.0-0.nightly-2023-11-25-110147' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-11-25-110147 True False 79m Cluster version is 4.15.0-0.nightly-2023-11-25-110147' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z '' ']' ++ KAFKA_USE_CUSTOM_PODSET=false ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/ui/render.sh ++ export SUITE_DIR=./tests/e2e/ui ++ SUITE_DIR=./tests/e2e/ui ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/ui ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test allinone + '[' 1 -ne 1 ']' + test_name=allinone + echo =========================================================================== =========================================================================== + info 'Rendering files for test allinone' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test allinone\e[0m' Rendering files for test allinone + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build + '[' _build '!=' _build ']' + mkdir -p allinone + cd allinone + export GET_URL_COMMAND + export URL + export JAEGER_NAME=all-in-one-ui + JAEGER_NAME=all-in-one-ui + '[' true = true ']' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./01-curl.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./04-test-ui-config.yaml + start_test production + '[' 1 -ne 1 ']' + test_name=production + echo =========================================================================== =========================================================================== + info 'Rendering files for test production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test production\e[0m' Rendering files for test production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build/allinone + '[' allinone '!=' _build ']' + cd .. + mkdir -p production + cd production + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + [[ true = true ]] + [[ true = true ]] + render_install_jaeger production-ui production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + '[' true = true ']' + INSECURE=true + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-forbbiden-access.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-curl.yaml + INSECURE=true + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./05-check-disabled-security.yaml + ASSERT_PRESENT=false + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./06-check-NO-gaID.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./08-check-gaID.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running ui E2E tests' Running ui E2E tests + cd tests/e2e/ui/_build + set +e + KUBECONFIG=/tmp/kubeconfig-4018005699 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 3 tests === RUN kuttl/harness === RUN kuttl/harness/allinone === PAUSE kuttl/harness/allinone === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/production === PAUSE kuttl/harness/production === CONT kuttl/harness/allinone logger.go:42: 15:15:23 | allinone | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 15:15:23 | allinone | Creating namespace: kuttl-test-casual-puma logger.go:42: 15:15:23 | allinone/0-install | starting test step 0-install logger.go:42: 15:15:23 | allinone/0-install | Jaeger:kuttl-test-casual-puma/all-in-one-ui created logger.go:42: 15:15:27 | allinone/0-install | test step completed 0-install logger.go:42: 15:15:27 | allinone/1-curl | starting test step 1-curl logger.go:42: 15:15:27 | allinone/1-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 15:15:27 | allinone/1-curl | Checking the Ingress host value was populated logger.go:42: 15:15:27 | allinone/1-curl | Try number 0 logger.go:42: 15:15:27 | allinone/1-curl | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 15:15:27 | allinone/1-curl | template was: logger.go:42: 15:15:27 | allinone/1-curl | {.items[0].status.ingress[0].host} logger.go:42: 15:15:27 | allinone/1-curl | object given to jsonpath engine was: logger.go:42: 15:15:27 | allinone/1-curl | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 15:15:27 | allinone/1-curl | logger.go:42: 15:15:27 | allinone/1-curl | logger.go:42: 15:15:37 | allinone/1-curl | Try number 1 logger.go:42: 15:15:38 | allinone/1-curl | Hostname is all-in-one-ui-kuttl-test-casual-puma.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com logger.go:42: 15:15:38 | allinone/1-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE all-in-one-ui] logger.go:42: 15:15:38 | allinone/1-curl | Checking an expected HTTP response logger.go:42: 15:15:38 | allinone/1-curl | Running in OpenShift logger.go:42: 15:15:38 | allinone/1-curl | User not provided. Getting the token... logger.go:42: 15:15:39 | allinone/1-curl | Warning: resource jaegers/all-in-one-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 15:15:45 | allinone/1-curl | Try number 1/30 the https://all-in-one-ui-kuttl-test-casual-puma.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 15:15:45 | allinone/1-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 15:15:45 | allinone/1-curl | Try number 2/30 the https://all-in-one-ui-kuttl-test-casual-puma.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 15:15:45 | allinone/1-curl | curl response asserted properly logger.go:42: 15:15:45 | allinone/1-curl | test step completed 1-curl logger.go:42: 15:15:45 | allinone/2-delete | starting test step 2-delete logger.go:42: 15:15:46 | allinone/2-delete | Jaeger:kuttl-test-casual-puma/all-in-one-ui created logger.go:42: 15:15:46 | allinone/2-delete | test step completed 2-delete logger.go:42: 15:15:46 | allinone/3-install | starting test step 3-install logger.go:42: 15:15:46 | allinone/3-install | Jaeger:kuttl-test-casual-puma/all-in-one-ui updated logger.go:42: 15:15:46 | allinone/3-install | test step completed 3-install logger.go:42: 15:15:46 | allinone/4-test-ui-config | starting test step 4-test-ui-config logger.go:42: 15:15:46 | allinone/4-test-ui-config | running command: [./ensure-ingress-host.sh] logger.go:42: 15:15:46 | allinone/4-test-ui-config | Checking the Ingress host value was populated logger.go:42: 15:15:46 | allinone/4-test-ui-config | Try number 0 logger.go:42: 15:15:46 | allinone/4-test-ui-config | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 15:15:46 | allinone/4-test-ui-config | template was: logger.go:42: 15:15:46 | allinone/4-test-ui-config | {.items[0].status.ingress[0].host} logger.go:42: 15:15:46 | allinone/4-test-ui-config | object given to jsonpath engine was: logger.go:42: 15:15:46 | allinone/4-test-ui-config | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 15:15:46 | allinone/4-test-ui-config | logger.go:42: 15:15:46 | allinone/4-test-ui-config | logger.go:42: 15:15:56 | allinone/4-test-ui-config | Try number 1 logger.go:42: 15:15:56 | allinone/4-test-ui-config | Hostname is all-in-one-ui-kuttl-test-casual-puma.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com logger.go:42: 15:15:56 | allinone/4-test-ui-config | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 15:15:57 | allinone/4-test-ui-config | time="2023-11-27T15:15:57Z" level=info msg="Querying https://all-in-one-ui-kuttl-test-casual-puma.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search..." logger.go:42: 15:15:57 | allinone/4-test-ui-config | time="2023-11-27T15:15:57Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 15:15:57 | allinone/4-test-ui-config | time="2023-11-27T15:15:57Z" level=info msg="Polling to https://all-in-one-ui-kuttl-test-casual-puma.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search" logger.go:42: 15:15:57 | allinone/4-test-ui-config | time="2023-11-27T15:15:57Z" level=info msg="Doing request number 0" logger.go:42: 15:15:57 | allinone/4-test-ui-config | time="2023-11-27T15:15:57Z" level=info msg="Content found and asserted!" logger.go:42: 15:15:57 | allinone/4-test-ui-config | time="2023-11-27T15:15:57Z" level=info msg="Success!" logger.go:42: 15:15:57 | allinone/4-test-ui-config | test step completed 4-test-ui-config logger.go:42: 15:15:57 | allinone | allinone events from ns kuttl-test-casual-puma: logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:27 +0000 UTC Normal Pod all-in-one-ui-844f6dbc5f-hnj6w Binding Scheduled Successfully assigned kuttl-test-casual-puma/all-in-one-ui-844f6dbc5f-hnj6w to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:27 +0000 UTC Normal Pod all-in-one-ui-844f6dbc5f-hnj6w AddedInterface Add eth0 [10.129.2.120/23] from ovn-kubernetes logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:27 +0000 UTC Normal Pod all-in-one-ui-844f6dbc5f-hnj6w.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:27 +0000 UTC Normal Pod all-in-one-ui-844f6dbc5f-hnj6w.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:27 +0000 UTC Normal Pod all-in-one-ui-844f6dbc5f-hnj6w.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:27 +0000 UTC Normal Pod all-in-one-ui-844f6dbc5f-hnj6w.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:27 +0000 UTC Normal Pod all-in-one-ui-844f6dbc5f-hnj6w.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:27 +0000 UTC Normal Pod all-in-one-ui-844f6dbc5f-hnj6w.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:27 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-844f6dbc5f SuccessfulCreate Created pod: all-in-one-ui-844f6dbc5f-hnj6w replicaset-controller logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:27 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-844f6dbc5f to 1 deployment-controller logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:40 +0000 UTC Normal Pod all-in-one-ui-844f6dbc5f-hnj6w.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:40 +0000 UTC Normal Pod all-in-one-ui-844f6dbc5f-hnj6w.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:40 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-844f6dbc5f SuccessfulDelete Deleted pod: all-in-one-ui-844f6dbc5f-hnj6w replicaset-controller logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:40 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-844f6dbc5f to 0 from 1 deployment-controller logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:41 +0000 UTC Normal Pod all-in-one-ui-75fd687c99-5p568 Binding Scheduled Successfully assigned kuttl-test-casual-puma/all-in-one-ui-75fd687c99-5p568 to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:41 +0000 UTC Normal Pod all-in-one-ui-75fd687c99-5p568 AddedInterface Add eth0 [10.128.2.67/23] from ovn-kubernetes logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:41 +0000 UTC Normal Pod all-in-one-ui-75fd687c99-5p568.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:41 +0000 UTC Normal Pod all-in-one-ui-75fd687c99-5p568.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:41 +0000 UTC Normal Pod all-in-one-ui-75fd687c99-5p568.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:41 +0000 UTC Normal Pod all-in-one-ui-75fd687c99-5p568.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:41 +0000 UTC Normal Pod all-in-one-ui-75fd687c99-5p568.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:41 +0000 UTC Normal Pod all-in-one-ui-75fd687c99-5p568.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:41 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-75fd687c99 SuccessfulCreate Created pod: all-in-one-ui-75fd687c99-5p568 replicaset-controller logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:41 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-75fd687c99 to 1 deployment-controller logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:46 +0000 UTC Normal Pod all-in-one-ui-75fd687c99-5p568.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:46 +0000 UTC Normal Pod all-in-one-ui-75fd687c99-5p568.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:49 +0000 UTC Normal Pod all-in-one-ui-7479dfd6fc-srs76 Binding Scheduled Successfully assigned kuttl-test-casual-puma/all-in-one-ui-7479dfd6fc-srs76 to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:49 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-7479dfd6fc SuccessfulCreate Created pod: all-in-one-ui-7479dfd6fc-srs76 replicaset-controller logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:49 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-7479dfd6fc to 1 deployment-controller logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:50 +0000 UTC Normal Pod all-in-one-ui-7479dfd6fc-srs76 AddedInterface Add eth0 [10.129.2.121/23] from ovn-kubernetes logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:50 +0000 UTC Normal Pod all-in-one-ui-7479dfd6fc-srs76.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:e724963dd365f319fdfaaa6159b16227b5744d8a6700974bdd9dfe4ddf40a580" already present on machine kubelet logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:50 +0000 UTC Normal Pod all-in-one-ui-7479dfd6fc-srs76.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 15:15:57 | allinone | 2023-11-27 15:15:50 +0000 UTC Normal Pod all-in-one-ui-7479dfd6fc-srs76.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 15:15:57 | allinone | Deleting namespace: kuttl-test-casual-puma === CONT kuttl/harness/production logger.go:42: 15:16:03 | production | Ignoring add-tracking-id.yaml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 15:16:03 | production | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 15:16:03 | production | Creating namespace: kuttl-test-maximum-dogfish logger.go:42: 15:16:03 | production/1-install | starting test step 1-install logger.go:42: 15:16:03 | production/1-install | Jaeger:kuttl-test-maximum-dogfish/production-ui created logger.go:42: 15:16:42 | production/1-install | test step completed 1-install logger.go:42: 15:16:42 | production/2-check-forbbiden-access | starting test step 2-check-forbbiden-access logger.go:42: 15:16:42 | production/2-check-forbbiden-access | running command: [./ensure-ingress-host.sh] logger.go:42: 15:16:42 | production/2-check-forbbiden-access | Checking the Ingress host value was populated logger.go:42: 15:16:42 | production/2-check-forbbiden-access | Try number 0 logger.go:42: 15:16:42 | production/2-check-forbbiden-access | Hostname is production-ui-kuttl-test-maximum-dogfish.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com logger.go:42: 15:16:42 | production/2-check-forbbiden-access | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE production-ui] logger.go:42: 15:16:42 | production/2-check-forbbiden-access | Checking an expected HTTP response logger.go:42: 15:16:42 | production/2-check-forbbiden-access | Running in OpenShift logger.go:42: 15:16:42 | production/2-check-forbbiden-access | Not using any secret logger.go:42: 15:16:42 | production/2-check-forbbiden-access | Try number 1/30 the https://production-ui-kuttl-test-maximum-dogfish.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 15:16:42 | production/2-check-forbbiden-access | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 15:16:42 | production/2-check-forbbiden-access | Try number 2/30 the https://production-ui-kuttl-test-maximum-dogfish.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 15:16:42 | production/2-check-forbbiden-access | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 15:16:52 | production/2-check-forbbiden-access | Try number 3/30 the https://production-ui-kuttl-test-maximum-dogfish.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 15:16:52 | production/2-check-forbbiden-access | curl response asserted properly logger.go:42: 15:16:52 | production/2-check-forbbiden-access | test step completed 2-check-forbbiden-access logger.go:42: 15:16:52 | production/3-curl | starting test step 3-curl logger.go:42: 15:16:52 | production/3-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 15:16:52 | production/3-curl | Checking the Ingress host value was populated logger.go:42: 15:16:52 | production/3-curl | Try number 0 logger.go:42: 15:16:52 | production/3-curl | Hostname is production-ui-kuttl-test-maximum-dogfish.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com logger.go:42: 15:16:52 | production/3-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 15:16:52 | production/3-curl | Checking an expected HTTP response logger.go:42: 15:16:52 | production/3-curl | Running in OpenShift logger.go:42: 15:16:52 | production/3-curl | User not provided. Getting the token... logger.go:42: 15:16:53 | production/3-curl | Warning: resource jaegers/production-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 15:17:00 | production/3-curl | Try number 1/30 the https://production-ui-kuttl-test-maximum-dogfish.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 15:17:00 | production/3-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 15:17:00 | production/3-curl | Try number 2/30 the https://production-ui-kuttl-test-maximum-dogfish.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 15:17:00 | production/3-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 15:17:10 | production/3-curl | Try number 3/30 the https://production-ui-kuttl-test-maximum-dogfish.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 15:17:10 | production/3-curl | curl response asserted properly logger.go:42: 15:17:10 | production/3-curl | test step completed 3-curl logger.go:42: 15:17:10 | production/4-install | starting test step 4-install logger.go:42: 15:17:10 | production/4-install | Jaeger:kuttl-test-maximum-dogfish/production-ui updated logger.go:42: 15:17:10 | production/4-install | test step completed 4-install logger.go:42: 15:17:10 | production/5-check-disabled-security | starting test step 5-check-disabled-security logger.go:42: 15:17:10 | production/5-check-disabled-security | running command: [./ensure-ingress-host.sh] logger.go:42: 15:17:10 | production/5-check-disabled-security | Checking the Ingress host value was populated logger.go:42: 15:17:10 | production/5-check-disabled-security | Try number 0 logger.go:42: 15:17:10 | production/5-check-disabled-security | Hostname is production-ui-kuttl-test-maximum-dogfish.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com logger.go:42: 15:17:10 | production/5-check-disabled-security | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 15:17:10 | production/5-check-disabled-security | Checking an expected HTTP response logger.go:42: 15:17:10 | production/5-check-disabled-security | Running in OpenShift logger.go:42: 15:17:10 | production/5-check-disabled-security | Not using any secret logger.go:42: 15:17:10 | production/5-check-disabled-security | Try number 1/30 the https://production-ui-kuttl-test-maximum-dogfish.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 15:17:10 | production/5-check-disabled-security | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 15:17:10 | production/5-check-disabled-security | Try number 2/30 the https://production-ui-kuttl-test-maximum-dogfish.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 15:17:10 | production/5-check-disabled-security | HTTP response is 403. 200 expected. Waiting 10 s logger.go:42: 15:17:20 | production/5-check-disabled-security | Try number 3/30 the https://production-ui-kuttl-test-maximum-dogfish.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 15:17:20 | production/5-check-disabled-security | curl response asserted properly logger.go:42: 15:17:20 | production/5-check-disabled-security | test step completed 5-check-disabled-security logger.go:42: 15:17:20 | production/6-check-NO-gaID | starting test step 6-check-NO-gaID logger.go:42: 15:17:20 | production/6-check-NO-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 15:17:20 | production/6-check-NO-gaID | Checking the Ingress host value was populated logger.go:42: 15:17:20 | production/6-check-NO-gaID | Try number 0 logger.go:42: 15:17:20 | production/6-check-NO-gaID | Hostname is production-ui-kuttl-test-maximum-dogfish.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com logger.go:42: 15:17:20 | production/6-check-NO-gaID | running command: [sh -c ASSERT_PRESENT=false EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 15:17:20 | production/6-check-NO-gaID | time="2023-11-27T15:17:20Z" level=info msg="Querying https://production-ui-kuttl-test-maximum-dogfish.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search..." logger.go:42: 15:17:20 | production/6-check-NO-gaID | time="2023-11-27T15:17:20Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 15:17:20 | production/6-check-NO-gaID | time="2023-11-27T15:17:20Z" level=info msg="Polling to https://production-ui-kuttl-test-maximum-dogfish.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search" logger.go:42: 15:17:20 | production/6-check-NO-gaID | time="2023-11-27T15:17:20Z" level=info msg="Doing request number 0" logger.go:42: 15:17:20 | production/6-check-NO-gaID | time="2023-11-27T15:17:20Z" level=info msg="Content not found and asserted it was not found!" logger.go:42: 15:17:20 | production/6-check-NO-gaID | time="2023-11-27T15:17:20Z" level=info msg="Success!" logger.go:42: 15:17:20 | production/6-check-NO-gaID | test step completed 6-check-NO-gaID logger.go:42: 15:17:20 | production/7-add-tracking-id | starting test step 7-add-tracking-id logger.go:42: 15:17:20 | production/7-add-tracking-id | running command: [sh -c kubectl apply -f add-tracking-id.yaml -n $NAMESPACE] logger.go:42: 15:17:21 | production/7-add-tracking-id | jaeger.jaegertracing.io/production-ui configured logger.go:42: 15:17:21 | production/7-add-tracking-id | test step completed 7-add-tracking-id logger.go:42: 15:17:21 | production/8-check-gaID | starting test step 8-check-gaID logger.go:42: 15:17:21 | production/8-check-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 15:17:21 | production/8-check-gaID | Checking the Ingress host value was populated logger.go:42: 15:17:21 | production/8-check-gaID | Try number 0 logger.go:42: 15:17:21 | production/8-check-gaID | Hostname is production-ui-kuttl-test-maximum-dogfish.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com logger.go:42: 15:17:21 | production/8-check-gaID | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 15:17:21 | production/8-check-gaID | time="2023-11-27T15:17:21Z" level=info msg="Querying https://production-ui-kuttl-test-maximum-dogfish.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search..." logger.go:42: 15:17:21 | production/8-check-gaID | time="2023-11-27T15:17:21Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 15:17:21 | production/8-check-gaID | time="2023-11-27T15:17:21Z" level=info msg="Polling to https://production-ui-kuttl-test-maximum-dogfish.apps.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com/search" logger.go:42: 15:17:21 | production/8-check-gaID | time="2023-11-27T15:17:21Z" level=info msg="Doing request number 0" logger.go:42: 15:17:21 | production/8-check-gaID | time="2023-11-27T15:17:21Z" level=warning msg="Found: false . Assert: true" logger.go:42: 15:17:21 | production/8-check-gaID | time="2023-11-27T15:17:21Z" level=warning msg="The condition of the test function was not accomplished" logger.go:42: 15:17:21 | production/8-check-gaID | time="2023-11-27T15:17:21Z" level=info msg="Doing request number 1" logger.go:42: 15:17:21 | production/8-check-gaID | time="2023-11-27T15:17:21Z" level=warning msg="Found: false . Assert: true" logger.go:42: 15:17:21 | production/8-check-gaID | time="2023-11-27T15:17:21Z" level=warning msg="The condition of the test function was not accomplished" logger.go:42: 15:17:29 | production/8-check-gaID | time="2023-11-27T15:17:29Z" level=info msg="Doing request number 2" logger.go:42: 15:17:29 | production/8-check-gaID | time="2023-11-27T15:17:29Z" level=info msg="Content found and asserted!" logger.go:42: 15:17:29 | production/8-check-gaID | time="2023-11-27T15:17:29Z" level=info msg="Success!" logger.go:42: 15:17:29 | production/8-check-gaID | test step completed 8-check-gaID logger.go:42: 15:17:29 | production | production events from ns kuttl-test-maximum-dogfish: logger.go:42: 15:17:29 | production | 2023-11-27 15:16:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaximumdogfishproductionui-1-952sqf5 Binding Scheduled Successfully assigned kuttl-test-maximum-dogfish/elasticsearch-cdm-kuttltestmaximumdogfishproductionui-1-952sqf5 to ip-10-0-30-134.ec2.internal default-scheduler logger.go:42: 15:17:29 | production | 2023-11-27 15:16:10 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestmaximumdogfishproductionui-1-95c4b865b SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestmaximumdogfishproductionui-1-952sqf5 replicaset-controller logger.go:42: 15:17:29 | production | 2023-11-27 15:16:10 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestmaximumdogfishproductionui-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestmaximumdogfishproductionui-1-95c4b865b to 1 deployment-controller logger.go:42: 15:17:29 | production | 2023-11-27 15:16:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaximumdogfishproductionui-1-952sqf5 AddedInterface Add eth0 [10.128.2.68/23] from ovn-kubernetes logger.go:42: 15:17:29 | production | 2023-11-27 15:16:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaximumdogfishproductionui-1-952sqf5.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:56ea62bfb0ca36e19a7b21aff3676e49511f05f72da5e76d6427fd8240f328a8" already present on machine kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaximumdogfishproductionui-1-952sqf5.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaximumdogfishproductionui-1-952sqf5.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaximumdogfishproductionui-1-952sqf5.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:79427bea6b5c37894b9782c3821d8b9074838e606daa4a743b2ae060856fa98a" already present on machine kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaximumdogfishproductionui-1-952sqf5.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaximumdogfishproductionui-1-952sqf5.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:21 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmaximumdogfishproductionui-1-952sqf5.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:26 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmaximumdogfishproductionui-1-952sqf5.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:37 +0000 UTC Normal Pod production-ui-collector-5b7c4bd9bb-r2zsf Binding Scheduled Successfully assigned kuttl-test-maximum-dogfish/production-ui-collector-5b7c4bd9bb-r2zsf to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 15:17:29 | production | 2023-11-27 15:16:37 +0000 UTC Normal ReplicaSet.apps production-ui-collector-5b7c4bd9bb SuccessfulCreate Created pod: production-ui-collector-5b7c4bd9bb-r2zsf replicaset-controller logger.go:42: 15:17:29 | production | 2023-11-27 15:16:37 +0000 UTC Normal Deployment.apps production-ui-collector ScalingReplicaSet Scaled up replica set production-ui-collector-5b7c4bd9bb to 1 deployment-controller logger.go:42: 15:17:29 | production | 2023-11-27 15:16:37 +0000 UTC Normal ReplicaSet.apps production-ui-query-d4bb9bd7d SuccessfulCreate Created pod: production-ui-query-d4bb9bd7d-f2d9f replicaset-controller logger.go:42: 15:17:29 | production | 2023-11-27 15:16:37 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-d4bb9bd7d to 1 deployment-controller logger.go:42: 15:17:29 | production | 2023-11-27 15:16:38 +0000 UTC Normal Pod production-ui-collector-5b7c4bd9bb-r2zsf AddedInterface Add eth0 [10.129.2.122/23] from ovn-kubernetes logger.go:42: 15:17:29 | production | 2023-11-27 15:16:38 +0000 UTC Normal Pod production-ui-collector-5b7c4bd9bb-r2zsf.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c12b97adae3dabdbb374df36766adf977dc61193def990d90c751c445d89d856" already present on machine kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:38 +0000 UTC Normal Pod production-ui-collector-5b7c4bd9bb-r2zsf.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:38 +0000 UTC Normal Pod production-ui-collector-5b7c4bd9bb-r2zsf.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:38 +0000 UTC Normal Pod production-ui-query-d4bb9bd7d-f2d9f Binding Scheduled Successfully assigned kuttl-test-maximum-dogfish/production-ui-query-d4bb9bd7d-f2d9f to ip-10-0-20-116.ec2.internal default-scheduler logger.go:42: 15:17:29 | production | 2023-11-27 15:16:39 +0000 UTC Warning Pod production-ui-query-d4bb9bd7d-f2d9f FailedMount MountVolume.SetUp failed for volume "production-ui-service-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:39 +0000 UTC Warning Pod production-ui-query-d4bb9bd7d-f2d9f FailedMount MountVolume.SetUp failed for volume "production-ui-ui-configuration-volume" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:39 +0000 UTC Normal Pod production-ui-query-d4bb9bd7d-f2d9f AddedInterface Add eth0 [10.129.2.123/23] from ovn-kubernetes logger.go:42: 15:17:29 | production | 2023-11-27 15:16:39 +0000 UTC Normal Pod production-ui-query-d4bb9bd7d-f2d9f.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:39 +0000 UTC Normal Pod production-ui-query-d4bb9bd7d-f2d9f.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:39 +0000 UTC Normal Pod production-ui-query-d4bb9bd7d-f2d9f.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:40 +0000 UTC Normal Pod production-ui-query-d4bb9bd7d-f2d9f.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:40 +0000 UTC Normal Pod production-ui-query-d4bb9bd7d-f2d9f.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:40 +0000 UTC Normal Pod production-ui-query-d4bb9bd7d-f2d9f.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:40 +0000 UTC Normal Pod production-ui-query-d4bb9bd7d-f2d9f.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:40 +0000 UTC Normal Pod production-ui-query-d4bb9bd7d-f2d9f.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:40 +0000 UTC Normal Pod production-ui-query-d4bb9bd7d-f2d9f.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:53 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 15:17:29 | production | 2023-11-27 15:16:53 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 15:17:29 | production | 2023-11-27 15:16:53 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 15:17:29 | production | 2023-11-27 15:16:55 +0000 UTC Normal Pod production-ui-query-d4bb9bd7d-f2d9f.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:55 +0000 UTC Normal Pod production-ui-query-d4bb9bd7d-f2d9f.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:55 +0000 UTC Normal Pod production-ui-query-d4bb9bd7d-f2d9f.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:55 +0000 UTC Normal ReplicaSet.apps production-ui-query-d4bb9bd7d SuccessfulDelete Deleted pod: production-ui-query-d4bb9bd7d-f2d9f replicaset-controller logger.go:42: 15:17:29 | production | 2023-11-27 15:16:55 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-d4bb9bd7d to 0 from 1 deployment-controller logger.go:42: 15:17:29 | production | 2023-11-27 15:16:56 +0000 UTC Normal Pod production-ui-query-7bfcbb9f95-xfspg Binding Scheduled Successfully assigned kuttl-test-maximum-dogfish/production-ui-query-7bfcbb9f95-xfspg to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 15:17:29 | production | 2023-11-27 15:16:56 +0000 UTC Normal Pod production-ui-query-7bfcbb9f95-xfspg AddedInterface Add eth0 [10.131.0.81/23] from ovn-kubernetes logger.go:42: 15:17:29 | production | 2023-11-27 15:16:56 +0000 UTC Normal Pod production-ui-query-7bfcbb9f95-xfspg.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:56 +0000 UTC Normal Pod production-ui-query-7bfcbb9f95-xfspg.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:56 +0000 UTC Normal Pod production-ui-query-7bfcbb9f95-xfspg.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:56 +0000 UTC Normal Pod production-ui-query-7bfcbb9f95-xfspg.spec.containers{oauth-proxy} Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a81a85db56907cbad4b0b581f1892982d63dee1f91cee4f82e62f6cc9a9df410" already present on machine kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:56 +0000 UTC Normal Pod production-ui-query-7bfcbb9f95-xfspg.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:56 +0000 UTC Normal Pod production-ui-query-7bfcbb9f95-xfspg.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:56 +0000 UTC Normal Pod production-ui-query-7bfcbb9f95-xfspg.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:56 +0000 UTC Normal ReplicaSet.apps production-ui-query-7bfcbb9f95 SuccessfulCreate Created pod: production-ui-query-7bfcbb9f95-xfspg replicaset-controller logger.go:42: 15:17:29 | production | 2023-11-27 15:16:56 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-7bfcbb9f95 to 1 deployment-controller logger.go:42: 15:17:29 | production | 2023-11-27 15:16:57 +0000 UTC Normal Pod production-ui-query-7bfcbb9f95-xfspg.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:16:57 +0000 UTC Normal Pod production-ui-query-7bfcbb9f95-xfspg.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:17:11 +0000 UTC Normal Pod production-ui-query-7bfcbb9f95-xfspg.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:17:11 +0000 UTC Normal Pod production-ui-query-7bfcbb9f95-xfspg.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:17:11 +0000 UTC Normal Pod production-ui-query-7bfcbb9f95-xfspg.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:17:11 +0000 UTC Normal ReplicaSet.apps production-ui-query-7bfcbb9f95 SuccessfulDelete Deleted pod: production-ui-query-7bfcbb9f95-xfspg replicaset-controller logger.go:42: 15:17:29 | production | 2023-11-27 15:17:11 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-7bfcbb9f95 to 0 from 1 deployment-controller logger.go:42: 15:17:29 | production | 2023-11-27 15:17:12 +0000 UTC Normal Pod production-ui-query-5d8f794956-rkvrd Binding Scheduled Successfully assigned kuttl-test-maximum-dogfish/production-ui-query-5d8f794956-rkvrd to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 15:17:29 | production | 2023-11-27 15:17:12 +0000 UTC Normal Pod production-ui-query-5d8f794956-rkvrd AddedInterface Add eth0 [10.131.0.82/23] from ovn-kubernetes logger.go:42: 15:17:29 | production | 2023-11-27 15:17:12 +0000 UTC Normal Pod production-ui-query-5d8f794956-rkvrd.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:17:12 +0000 UTC Normal Pod production-ui-query-5d8f794956-rkvrd.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:17:12 +0000 UTC Normal Pod production-ui-query-5d8f794956-rkvrd.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:17:12 +0000 UTC Normal Pod production-ui-query-5d8f794956-rkvrd.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:17:12 +0000 UTC Normal ReplicaSet.apps production-ui-query-5d8f794956 SuccessfulCreate Created pod: production-ui-query-5d8f794956-rkvrd replicaset-controller logger.go:42: 15:17:29 | production | 2023-11-27 15:17:12 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-5d8f794956 to 1 deployment-controller logger.go:42: 15:17:29 | production | 2023-11-27 15:17:13 +0000 UTC Normal Pod production-ui-query-5d8f794956-rkvrd.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:17:13 +0000 UTC Normal Pod production-ui-query-5d8f794956-rkvrd.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:17:22 +0000 UTC Normal Pod production-ui-query-5d8f794956-rkvrd.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:17:22 +0000 UTC Normal Pod production-ui-query-5d8f794956-rkvrd.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:17:22 +0000 UTC Normal ReplicaSet.apps production-ui-query-5d8f794956 SuccessfulDelete Deleted pod: production-ui-query-5d8f794956-rkvrd replicaset-controller logger.go:42: 15:17:29 | production | 2023-11-27 15:17:22 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-5d8f794956 to 0 from 1 deployment-controller logger.go:42: 15:17:29 | production | 2023-11-27 15:17:23 +0000 UTC Normal Pod production-ui-query-74f58bf9b9-rvlm4 Binding Scheduled Successfully assigned kuttl-test-maximum-dogfish/production-ui-query-74f58bf9b9-rvlm4 to ip-10-0-124-246.ec2.internal default-scheduler logger.go:42: 15:17:29 | production | 2023-11-27 15:17:23 +0000 UTC Normal Pod production-ui-query-74f58bf9b9-rvlm4 AddedInterface Add eth0 [10.131.0.83/23] from ovn-kubernetes logger.go:42: 15:17:29 | production | 2023-11-27 15:17:23 +0000 UTC Normal Pod production-ui-query-74f58bf9b9-rvlm4.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a07706a058469605ebbfaef01660e2b77ad064721615a2e37bb5c7c6aad1156f" already present on machine kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:17:23 +0000 UTC Normal Pod production-ui-query-74f58bf9b9-rvlm4.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:17:23 +0000 UTC Normal Pod production-ui-query-74f58bf9b9-rvlm4.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:17:23 +0000 UTC Normal Pod production-ui-query-74f58bf9b9-rvlm4.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:b99b1666405ba55c81bdc1f4cb385228fd513f94f7b74881b78e3cb7f63da937" already present on machine kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:17:23 +0000 UTC Normal ReplicaSet.apps production-ui-query-74f58bf9b9 SuccessfulCreate Created pod: production-ui-query-74f58bf9b9-rvlm4 replicaset-controller logger.go:42: 15:17:29 | production | 2023-11-27 15:17:23 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-74f58bf9b9 to 1 deployment-controller logger.go:42: 15:17:29 | production | 2023-11-27 15:17:24 +0000 UTC Normal Pod production-ui-query-74f58bf9b9-rvlm4.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 15:17:29 | production | 2023-11-27 15:17:24 +0000 UTC Normal Pod production-ui-query-74f58bf9b9-rvlm4.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 15:17:29 | production | Deleting namespace: kuttl-test-maximum-dogfish === CONT kuttl/harness/artifacts logger.go:42: 15:17:35 | artifacts | Creating namespace: kuttl-test-present-buck logger.go:42: 15:17:35 | artifacts | artifacts events from ns kuttl-test-present-buck: logger.go:42: 15:17:35 | artifacts | Deleting namespace: kuttl-test-present-buck === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (137.74s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/allinone (39.97s) --- PASS: kuttl/harness/production (92.07s) --- PASS: kuttl/harness/artifacts (5.66s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name ui --report --output /logs/artifacts/ui.xml ./artifacts/kuttl-report.xml time="2023-11-27T15:17:41Z" level=debug msg="Setting a new name for the test suites" time="2023-11-27T15:17:41Z" level=debug msg="Removing 'artifacts' TestCase" time="2023-11-27T15:17:41Z" level=debug msg="normalizing test case names" time="2023-11-27T15:17:41Z" level=debug msg="ui/allinone -> ui_allinone" time="2023-11-27T15:17:41Z" level=debug msg="ui/production -> ui_production" time="2023-11-27T15:17:41Z" level=debug msg="ui/artifacts -> ui_artifacts" +---------------+--------+ | NAME | RESULT | +---------------+--------+ | ui_allinone | passed | | ui_production | passed | | ui_artifacts | passed | +---------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh upgrade false true + '[' 3 -ne 3 ']' + test_suite_name=upgrade + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/upgrade.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-upgrade make[2]: Entering directory '/tmp/jaeger-tests' make docker JAEGER_VERSION=1.51.1 IMG="quay.io//jaeger-operator:next" make[3]: Entering directory '/tmp/jaeger-tests' [ ! -z "true" ] || docker build --build-arg=GOPROXY= --build-arg=VERSION="1.51.0" --build-arg=JAEGER_VERSION=1.51.1 --build-arg=TARGETARCH= --build-arg VERSION_DATE=2023-11-27T15:17:41Z --build-arg VERSION_PKG="github.com/jaegertracing/jaeger-operator/pkg/version" -t "quay.io//jaeger-operator:next" . make[3]: Leaving directory '/tmp/jaeger-tests' touch build-e2e-upgrade-image SKIP_ES_EXTERNAL=true IMG=quay.io//jaeger-operator:"1.51.0" JAEGER_OPERATOR_VERSION="1.51.0" JAEGER_VERSION="1.51.0" ./tests/e2e/upgrade/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-11-25-110147 True False 81m Cluster version is 4.15.0-0.nightly-2023-11-25-110147' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2023-11-25-110147 True False 81m Cluster version is 4.15.0-0.nightly-2023-11-25-110147' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z '' ']' ++ KAFKA_USE_CUSTOM_PODSET=false ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/upgrade/render.sh ++ export SUITE_DIR=./tests/e2e/upgrade ++ SUITE_DIR=./tests/e2e/upgrade ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/upgrade ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + export JAEGER_NAME + '[' true = true ']' + skip_test upgrade 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade + warning 'upgrade: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade: Test not supported in OpenShift\e[0m' WAR: upgrade: Test not supported in OpenShift + '[' true = true ']' + skip_test upgrade-from-latest-release 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade-from-latest-release + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade-from-latest-release + warning 'upgrade-from-latest-release: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade-from-latest-release: Test not supported in OpenShift\e[0m' WAR: upgrade-from-latest-release: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running upgrade E2E tests' Running upgrade E2E tests + cd tests/e2e/upgrade/_build + set +e + KUBECONFIG=/tmp/kubeconfig-4018005699 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-1i9j6n46-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 15:17:43 | artifacts | Creating namespace: kuttl-test-smiling-caribou logger.go:42: 15:17:43 | artifacts | artifacts events from ns kuttl-test-smiling-caribou: logger.go:42: 15:17:43 | artifacts | Deleting namespace: kuttl-test-smiling-caribou === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.00s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.96s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name upgrade --report --output /logs/artifacts/upgrade.xml ./artifacts/kuttl-report.xml time="2023-11-27T15:17:49Z" level=debug msg="Setting a new name for the test suites" time="2023-11-27T15:17:49Z" level=debug msg="Removing 'artifacts' TestCase" time="2023-11-27T15:17:49Z" level=debug msg="normalizing test case names" time="2023-11-27T15:17:49Z" level=debug msg="upgrade/artifacts -> upgrade_artifacts" +-------------------+--------+ | NAME | RESULT | +-------------------+--------+ | upgrade_artifacts | passed | +-------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests'