% Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 3831 100 3831 0 0 31448 0 --:--:-- --:--:-- --:--:-- 31661 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 1953 100 1953 0 0 22075 0 --:--:-- --:--:-- --:--:-- 21943 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 350 100 350 0 0 3281 0 --:--:-- --:--:-- --:--:-- 3301 Installing kuttl Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/kubectl-kuttl https://github.com/kudobuilder/kuttl/releases/download/v0.15.0/kubectl-kuttl_0.15.0_linux_x86_64 KUBECONFIG file is: /tmp/kubeconfig-2576283944 for suite in elasticsearch examples generate miscellaneous sidecar streaming ui upgrade; do \ make run-e2e-tests-$suite ; \ done make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh elasticsearch false true + '[' 3 -ne 3 ']' + test_suite_name=elasticsearch + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/elasticsearch.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-elasticsearch make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true \ KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ ./tests/e2e/elasticsearch/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-18-050837 True False 9m52s Cluster version is 4.15.0-0.nightly-2024-01-18-050837' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-18-050837 True False 9m52s Cluster version is 4.15.0-0.nightly-2024-01-18-050837' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/elasticsearch/render.sh ++ export SUITE_DIR=./tests/e2e/elasticsearch ++ SUITE_DIR=./tests/e2e/elasticsearch ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/elasticsearch ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + start_test es-from-aio-to-production + '[' 1 -ne 1 ']' + test_name=es-from-aio-to-production + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-from-aio-to-production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-from-aio-to-production\e[0m' Rendering files for test es-from-aio-to-production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-from-aio-to-production + cd es-from-aio-to-production + jaeger_name=my-jaeger + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 03 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=03 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./03-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch.redundancyPolicy="ZeroRedundancy"' ./03-install.yaml + render_smoke_test my-jaeger true 04 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=04 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./04-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test es-increasing-replicas + '[' 1 -ne 1 ']' + test_name=es-increasing-replicas + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-increasing-replicas' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-increasing-replicas\e[0m' Rendering files for test es-increasing-replicas + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-from-aio-to-production + '[' es-from-aio-to-production '!=' _build ']' + cd .. + mkdir -p es-increasing-replicas + cd es-increasing-replicas + jaeger_name=simple-prod + '[' true = true ']' + jaeger_deployment_mode=production_autoprovisioned + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.replicas=2 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.query.replicas=2 ./02-install.yaml + cp ./01-assert.yaml ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.replicas=2 ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .status.readyReplicas=2 ./02-assert.yaml + render_smoke_test simple-prod true 03 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=03 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./03-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + cp ./02-install.yaml ./04-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.elasticsearch.nodeCount=2 ./04-install.yaml + /tmp/jaeger-tests/bin/gomplate -f ./openshift-check-es-nodes.yaml.template -o ./05-check-es-nodes.yaml + '[' true = true ']' + skip_test es-index-cleaner-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-increasing-replicas + '[' es-increasing-replicas '!=' _build ']' + cd .. + rm -rf es-index-cleaner-upstream + warning 'es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_index_cleaner -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-index-cleaner-autoprov + '[' 1 -ne 1 ']' + test_name=es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-index-cleaner-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-index-cleaner-autoprov\e[0m' Rendering files for test es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-index-cleaner-autoprov + cd es-index-cleaner-autoprov + jaeger_name=test-es-index-cleaner-with-prefix + cronjob_name=test-es-index-cleaner-with-prefix-es-index-cleaner + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + cp ../../es-index-cleaner-upstream/04-assert.yaml ../../es-index-cleaner-upstream/README.md . + render_install_jaeger test-es-index-cleaner-with-prefix production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options.es.index-prefix=""' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.enabled=false ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.numberOfDays=0 ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.esIndexCleaner.schedule="*/1 * * * *"' ./01-install.yaml + render_report_spans test-es-index-cleaner-with-prefix true 5 00 true 02 + '[' 6 -ne 6 ']' + jaeger=test-es-index-cleaner-with-prefix + is_secured=true + number_of_spans=5 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + export JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=5 + DAYS=5 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + sed 's~enabled: false~enabled: true~gi' ./01-install.yaml + CRONJOB_NAME=test-es-index-cleaner-with-prefix-es-index-cleaner + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./04-wait-es-index-cleaner.yaml + /tmp/jaeger-tests/bin/gomplate -f ./01-install.yaml -o ./05-install.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 00 06 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=00 + test_step=06 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=test-es-index-cleaner-with-prefix-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=test-es-index-cleaner-with-prefix-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./06-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./06-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.2"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.2"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.2"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.2 ++ version_ge 5.8.2 5.4 +++ echo 5.8.2 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.2 == 5.8.2 + '[' -n '' ']' + skip_test es-index-cleaner-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-index-cleaner-autoprov + '[' es-index-cleaner-autoprov '!=' _build ']' + cd .. + rm -rf es-index-cleaner-managed + warning 'es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + start_test es-multiinstance + '[' 1 -ne 1 ']' + test_name=es-multiinstance + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-multiinstance' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-multiinstance\e[0m' Rendering files for test es-multiinstance + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-multiinstance + cd es-multiinstance + jaeger_name=instance-1 + render_install_jaeger instance-1 production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=instance-1 + JAEGER_NAME=instance-1 + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f ./03-create-second-instance.yaml.template -o 03-create-second-instance.yaml + '[' true = true ']' + skip_test es-rollover-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-rollover-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-multiinstance + '[' es-multiinstance '!=' _build ']' + cd .. + rm -rf es-rollover-upstream + warning 'es-rollover-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_rollover -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-rollover-autoprov + '[' 1 -ne 1 ']' + test_name=es-rollover-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-rollover-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-rollover-autoprov\e[0m' Rendering files for test es-rollover-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-rollover-autoprov + cd es-rollover-autoprov + cp ../../es-rollover-upstream/05-assert.yaml ../../es-rollover-upstream/05-install.yaml ../../es-rollover-upstream/README.md . + jaeger_name=my-jaeger + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_report_spans my-jaeger true 2 00 true 02 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 00 03 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=00 + test_step=03 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./03-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./03-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 01 04 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=01 + test_step=04 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=01 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./04-check-indices.yaml + JOB_NUMBER=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./04-assert.yaml + render_report_spans my-jaeger true 2 02 true 06 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=02 + ensure_reported_spans=true + test_step=06 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=02 + JOB_NUMBER=02 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./06-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./06-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 02 07 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=02 + test_step=07 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=02 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./07-check-indices.yaml + JOB_NUMBER=02 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./07-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' 03 08 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + job_number=03 + test_step=08 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=03 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./08-check-indices.yaml + JOB_NUMBER=03 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./08-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' 04 09 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + job_number=04 + test_step=09 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=04 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./09-check-indices.yaml + JOB_NUMBER=04 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./09-assert.yaml + render_report_spans my-jaeger true 2 03 true 10 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=03 + ensure_reported_spans=true + test_step=10 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=03 + JOB_NUMBER=03 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./10-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./10-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + CRONJOB_NAME=my-jaeger-es-rollover + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./11-wait-rollover.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-000002'\'',' 05 11 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-000002'\'',' + job_number=05 + test_step=11 + escape_command ''\''--name'\'', '\''jaeger-span-000002'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-000002'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-000002'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-000002'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=05 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./11-check-indices.yaml + JOB_NUMBER=05 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./11-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' 06 12 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + job_number=06 + test_step=12 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=06 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./12-check-indices.yaml + JOB_NUMBER=06 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./12-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.2"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.2"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.2"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.2 ++ version_ge 5.8.2 5.4 +++ echo 5.8.2 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.2 == 5.8.2 + '[' -n '' ']' + skip_test es-rollover-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-rollover-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-rollover-autoprov + '[' es-rollover-autoprov '!=' _build ']' + cd .. + rm -rf es-rollover-managed + warning 'es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + skip_test es-spark-dependencies 'This test is not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=es-spark-dependencies + message='This test is not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + rm -rf es-spark-dependencies + warning 'es-spark-dependencies: This test is not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-spark-dependencies: This test is not supported in OpenShift\e[0m' WAR: es-spark-dependencies: This test is not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running elasticsearch E2E tests' Running elasticsearch E2E tests + cd tests/e2e/elasticsearch/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2576283944 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/es-from-aio-to-production === PAUSE kuttl/harness/es-from-aio-to-production === RUN kuttl/harness/es-increasing-replicas === PAUSE kuttl/harness/es-increasing-replicas === RUN kuttl/harness/es-index-cleaner-autoprov === PAUSE kuttl/harness/es-index-cleaner-autoprov === RUN kuttl/harness/es-multiinstance === PAUSE kuttl/harness/es-multiinstance === RUN kuttl/harness/es-rollover-autoprov === PAUSE kuttl/harness/es-rollover-autoprov === RUN kuttl/harness/es-simple-prod === PAUSE kuttl/harness/es-simple-prod === CONT kuttl/harness/artifacts logger.go:42: 06:49:15 | artifacts | Creating namespace: kuttl-test-assuring-tarpon logger.go:42: 06:49:15 | artifacts | artifacts events from ns kuttl-test-assuring-tarpon: logger.go:42: 06:49:15 | artifacts | Deleting namespace: kuttl-test-assuring-tarpon === CONT kuttl/harness/es-multiinstance logger.go:42: 06:49:21 | es-multiinstance | Ignoring 03-create-second-instance.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 06:49:21 | es-multiinstance | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 06:49:21 | es-multiinstance | Creating namespace: kuttl-test-full-lioness logger.go:42: 06:49:21 | es-multiinstance/0-clear-namespace | starting test step 0-clear-namespace logger.go:42: 06:49:21 | es-multiinstance/0-clear-namespace | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --ignore-not-found=true] logger.go:42: 06:49:21 | es-multiinstance/0-clear-namespace | test step completed 0-clear-namespace logger.go:42: 06:49:21 | es-multiinstance/1-install | starting test step 1-install logger.go:42: 06:49:21 | es-multiinstance/1-install | Jaeger:kuttl-test-full-lioness/instance-1 created logger.go:42: 06:50:14 | es-multiinstance/1-install | test step completed 1-install logger.go:42: 06:50:14 | es-multiinstance/2-create-namespace | starting test step 2-create-namespace logger.go:42: 06:50:14 | es-multiinstance/2-create-namespace | running command: [sh -c kubectl create namespace jaeger-e2e-multiinstance-test] logger.go:42: 06:50:14 | es-multiinstance/2-create-namespace | namespace/jaeger-e2e-multiinstance-test created logger.go:42: 06:50:14 | es-multiinstance/2-create-namespace | test step completed 2-create-namespace logger.go:42: 06:50:14 | es-multiinstance/3-create-second-instance | starting test step 3-create-second-instance logger.go:42: 06:50:14 | es-multiinstance/3-create-second-instance | running command: [sh -c kubectl apply -f ./01-install.yaml -n jaeger-e2e-multiinstance-test] logger.go:42: 06:50:16 | es-multiinstance/3-create-second-instance | jaeger.jaegertracing.io/instance-1 created logger.go:42: 06:50:16 | es-multiinstance/3-create-second-instance | running command: [sh -c /tmp/jaeger-tests/bin/kubectl-kuttl assert ./01-assert.yaml -n jaeger-e2e-multiinstance-test --timeout 1000] logger.go:42: 06:51:06 | es-multiinstance/3-create-second-instance | assert is valid logger.go:42: 06:51:06 | es-multiinstance/3-create-second-instance | test step completed 3-create-second-instance logger.go:42: 06:51:06 | es-multiinstance/4-check-secrets | starting test step 4-check-secrets logger.go:42: 06:51:06 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n $NAMESPACE > secret1] logger.go:42: 06:51:07 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n jaeger-e2e-multiinstance-test > secret2] logger.go:42: 06:51:07 | es-multiinstance/4-check-secrets | running command: [sh -c cmp --silent secret1 secret2 || exit 0] logger.go:42: 06:51:07 | es-multiinstance/4-check-secrets | test step completed 4-check-secrets logger.go:42: 06:51:07 | es-multiinstance/5-delete | starting test step 5-delete logger.go:42: 06:51:07 | es-multiinstance/5-delete | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --wait=false] logger.go:42: 06:51:07 | es-multiinstance/5-delete | namespace "jaeger-e2e-multiinstance-test" deleted logger.go:42: 06:51:07 | es-multiinstance/5-delete | test step completed 5-delete logger.go:42: 06:51:07 | es-multiinstance | es-multiinstance events from ns kuttl-test-full-lioness: logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:28 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestfulllionessinstance1-1-84b85fd679 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestfulllionessinstance1-1-84b85fd6fr2j8 replicaset-controller logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfulllionessinstance1-1-84b85fd6fr2j8 Binding Scheduled Successfully assigned kuttl-test-full-lioness/elasticsearch-cdm-kuttltestfulllionessinstance1-1-84b85fd6fr2j8 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:28 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestfulllionessinstance1-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestfulllionessinstance1-1-84b85fd679 to 1 deployment-controller logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:28 +0000 UTC Normal Pod instance-1-collector-596c774b64-f7vrh Binding Scheduled Successfully assigned kuttl-test-full-lioness/instance-1-collector-596c774b64-f7vrh to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:28 +0000 UTC Normal ReplicaSet.apps instance-1-collector-596c774b64 SuccessfulCreate Created pod: instance-1-collector-596c774b64-f7vrh replicaset-controller logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:28 +0000 UTC Normal Deployment.apps instance-1-collector ScalingReplicaSet Scaled up replica set instance-1-collector-596c774b64 to 1 deployment-controller logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:28 +0000 UTC Normal Pod instance-1-query-5db4454647-8z9z6 Binding Scheduled Successfully assigned kuttl-test-full-lioness/instance-1-query-5db4454647-8z9z6 to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:28 +0000 UTC Normal ReplicaSet.apps instance-1-query-5db4454647 SuccessfulCreate Created pod: instance-1-query-5db4454647-8z9z6 replicaset-controller logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:28 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-5db4454647 to 1 deployment-controller logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfulllionessinstance1-1-84b85fd6fr2j8 AddedInterface Add eth0 [10.131.0.18/23] from ovn-kubernetes logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfulllionessinstance1-1-84b85fd6fr2j8.spec.containers{elasticsearch} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:3ad088cbfcfb006d87762e96a70ff989e75d49872954347f7efe34d9529c1b7a" kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:29 +0000 UTC Normal Pod instance-1-collector-596c774b64-f7vrh AddedInterface Add eth0 [10.128.2.28/23] from ovn-kubernetes logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:29 +0000 UTC Normal Pod instance-1-collector-596c774b64-f7vrh.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:29 +0000 UTC Normal Pod instance-1-query-5db4454647-8z9z6 AddedInterface Add eth0 [10.129.2.16/23] from ovn-kubernetes logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:29 +0000 UTC Normal Pod instance-1-query-5db4454647-8z9z6.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:33 +0000 UTC Normal Pod instance-1-collector-596c774b64-f7vrh.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" in 4.226s (4.226s including waiting) kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:33 +0000 UTC Normal Pod instance-1-collector-596c774b64-f7vrh.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:33 +0000 UTC Normal Pod instance-1-collector-596c774b64-f7vrh.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:34 +0000 UTC Warning Pod instance-1-collector-596c774b64-f7vrh.spec.containers{jaeger-collector} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfulllionessinstance1-1-84b85fd6fr2j8.spec.containers{elasticsearch} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:3ad088cbfcfb006d87762e96a70ff989e75d49872954347f7efe34d9529c1b7a" in 6.189s (6.189s including waiting) kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfulllionessinstance1-1-84b85fd6fr2j8.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfulllionessinstance1-1-84b85fd6fr2j8.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfulllionessinstance1-1-84b85fd6fr2j8.spec.containers{proxy} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:911ee3f26ae71eab71bf2a1455d1c2ea871cddc78a51189e3ad2f18f63bd3d91" kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:36 +0000 UTC Normal Pod instance-1-query-5db4454647-8z9z6.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" in 6.946s (6.946s including waiting) kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:36 +0000 UTC Normal Pod instance-1-query-5db4454647-8z9z6.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:36 +0000 UTC Normal Pod instance-1-query-5db4454647-8z9z6.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:36 +0000 UTC Normal Pod instance-1-query-5db4454647-8z9z6.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfulllionessinstance1-1-84b85fd6fr2j8.spec.containers{proxy} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:911ee3f26ae71eab71bf2a1455d1c2ea871cddc78a51189e3ad2f18f63bd3d91" in 3.173s (3.173s including waiting) kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfulllionessinstance1-1-84b85fd6fr2j8.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfulllionessinstance1-1-84b85fd6fr2j8.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:38 +0000 UTC Warning Pod instance-1-collector-596c774b64-f7vrh.spec.containers{jaeger-collector} Unhealthy Readiness probe failed: Get "http://10.128.2.28:14269/": dial tcp 10.128.2.28:14269: connect: connection refused kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:38 +0000 UTC Normal Pod instance-1-collector-596c774b64-f7vrh.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:41 +0000 UTC Normal Pod instance-1-query-5db4454647-8z9z6.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" in 5.416s (5.416s including waiting) kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:42 +0000 UTC Normal Pod instance-1-query-5db4454647-8z9z6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:42 +0000 UTC Normal Pod instance-1-query-5db4454647-8z9z6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:42 +0000 UTC Normal Pod instance-1-query-5db4454647-8z9z6.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:44 +0000 UTC Warning Pod instance-1-collector-596c774b64-f7vrh.spec.containers{jaeger-collector} BackOff Back-off restarting failed container jaeger-collector in pod instance-1-collector-596c774b64-f7vrh_kuttl-test-full-lioness(09d6a38e-407e-4794-8616-3654d9e78cdc) kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:45 +0000 UTC Normal Pod instance-1-query-5db4454647-8z9z6.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" in 3.336s (3.336s including waiting) kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:45 +0000 UTC Normal Pod instance-1-query-5db4454647-8z9z6.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:45 +0000 UTC Normal Pod instance-1-query-5db4454647-8z9z6.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:45 +0000 UTC Normal Pod instance-1-query-5db4454647-8z9z6.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:46 +0000 UTC Warning Pod instance-1-query-5db4454647-8z9z6.spec.containers{jaeger-query} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:49 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfulllionessinstance1-1-84b85fd6fr2j8.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:49:51 +0000 UTC Warning Pod instance-1-query-5db4454647-8z9z6.spec.containers{jaeger-query} BackOff Back-off restarting failed container jaeger-query in pod instance-1-query-5db4454647-8z9z6_kuttl-test-full-lioness(94de412a-fe95-42d2-8f89-0b71540f9491) kubelet logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:50:30 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:50:32 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:50:32 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:51:02 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:51:02 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod instance-1-collector-596c774b64-f7vrh horizontal-pod-autoscaler logger.go:42: 06:51:07 | es-multiinstance | 2024-01-22 06:51:02 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 06:51:07 | es-multiinstance | Deleting namespace: kuttl-test-full-lioness === CONT kuttl/harness/es-simple-prod logger.go:42: 06:51:15 | es-simple-prod | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 06:51:15 | es-simple-prod | Creating namespace: kuttl-test-rational-alien logger.go:42: 06:51:15 | es-simple-prod | es-simple-prod events from ns kuttl-test-rational-alien: logger.go:42: 06:51:15 | es-simple-prod | Deleting namespace: kuttl-test-rational-alien === CONT kuttl/harness/es-rollover-autoprov logger.go:42: 06:51:21 | es-rollover-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 06:51:21 | es-rollover-autoprov | Creating namespace: kuttl-test-busy-turkey logger.go:42: 06:51:21 | es-rollover-autoprov/1-install | starting test step 1-install logger.go:42: 06:51:21 | es-rollover-autoprov/1-install | Jaeger:kuttl-test-busy-turkey/my-jaeger created logger.go:42: 06:51:56 | es-rollover-autoprov/1-install | test step completed 1-install logger.go:42: 06:51:56 | es-rollover-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 06:51:56 | es-rollover-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 06:51:59 | es-rollover-autoprov/2-report-spans | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 06:52:06 | es-rollover-autoprov/2-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 06:52:06 | es-rollover-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 06:52:07 | es-rollover-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 06:52:33 | es-rollover-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 06:52:33 | es-rollover-autoprov/3-check-indices | starting test step 3-check-indices logger.go:42: 06:52:33 | es-rollover-autoprov/3-check-indices | Job:kuttl-test-busy-turkey/00-check-indices created logger.go:42: 06:52:37 | es-rollover-autoprov/3-check-indices | test step completed 3-check-indices logger.go:42: 06:52:37 | es-rollover-autoprov/4-check-indices | starting test step 4-check-indices logger.go:42: 06:52:37 | es-rollover-autoprov/4-check-indices | Job:kuttl-test-busy-turkey/01-check-indices created logger.go:42: 06:52:41 | es-rollover-autoprov/4-check-indices | test step completed 4-check-indices logger.go:42: 06:52:41 | es-rollover-autoprov/5-install | starting test step 5-install logger.go:42: 06:52:41 | es-rollover-autoprov/5-install | Jaeger:kuttl-test-busy-turkey/my-jaeger updated logger.go:42: 06:52:50 | es-rollover-autoprov/5-install | test step completed 5-install logger.go:42: 06:52:50 | es-rollover-autoprov/6-report-spans | starting test step 6-report-spans logger.go:42: 06:52:50 | es-rollover-autoprov/6-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 06:52:59 | es-rollover-autoprov/6-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JOB_NUMBER=02 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-02-job.yaml] logger.go:42: 06:53:00 | es-rollover-autoprov/6-report-spans | running command: [sh -c kubectl apply -f report-span-02-job.yaml -n $NAMESPACE] logger.go:42: 06:53:00 | es-rollover-autoprov/6-report-spans | job.batch/02-report-span created logger.go:42: 06:53:25 | es-rollover-autoprov/6-report-spans | test step completed 6-report-spans logger.go:42: 06:53:25 | es-rollover-autoprov/7-check-indices | starting test step 7-check-indices logger.go:42: 06:53:25 | es-rollover-autoprov/7-check-indices | Job:kuttl-test-busy-turkey/02-check-indices created logger.go:42: 06:53:29 | es-rollover-autoprov/7-check-indices | test step completed 7-check-indices logger.go:42: 06:53:29 | es-rollover-autoprov/8-check-indices | starting test step 8-check-indices logger.go:42: 06:53:29 | es-rollover-autoprov/8-check-indices | Job:kuttl-test-busy-turkey/03-check-indices created logger.go:42: 06:53:33 | es-rollover-autoprov/8-check-indices | test step completed 8-check-indices logger.go:42: 06:53:33 | es-rollover-autoprov/9-check-indices | starting test step 9-check-indices logger.go:42: 06:53:33 | es-rollover-autoprov/9-check-indices | Job:kuttl-test-busy-turkey/04-check-indices created logger.go:42: 06:53:37 | es-rollover-autoprov/9-check-indices | test step completed 9-check-indices logger.go:42: 06:53:37 | es-rollover-autoprov/10-report-spans | starting test step 10-report-spans logger.go:42: 06:53:37 | es-rollover-autoprov/10-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 06:53:46 | es-rollover-autoprov/10-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JOB_NUMBER=03 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-03-job.yaml] logger.go:42: 06:53:47 | es-rollover-autoprov/10-report-spans | running command: [sh -c kubectl apply -f report-span-03-job.yaml -n $NAMESPACE] logger.go:42: 06:53:47 | es-rollover-autoprov/10-report-spans | job.batch/03-report-span created logger.go:42: 06:54:11 | es-rollover-autoprov/10-report-spans | test step completed 10-report-spans logger.go:42: 06:54:11 | es-rollover-autoprov/11-check-indices | starting test step 11-check-indices logger.go:42: 06:54:11 | es-rollover-autoprov/11-check-indices | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob my-jaeger-es-rollover --namespace $NAMESPACE] logger.go:42: 06:54:23 | es-rollover-autoprov/11-check-indices | time="2024-01-22T06:54:23Z" level=debug msg="Checking if the my-jaeger-es-rollover CronJob exists" logger.go:42: 06:54:23 | es-rollover-autoprov/11-check-indices | time="2024-01-22T06:54:23Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 06:54:23 | es-rollover-autoprov/11-check-indices | time="2024-01-22T06:54:23Z" level=info msg="Cronjob my-jaeger-es-rollover found successfully" logger.go:42: 06:54:23 | es-rollover-autoprov/11-check-indices | time="2024-01-22T06:54:23Z" level=debug msg="Waiting for the next scheduled job from my-jaeger-es-rollover cronjob" logger.go:42: 06:54:23 | es-rollover-autoprov/11-check-indices | time="2024-01-22T06:54:23Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 06:54:33 | es-rollover-autoprov/11-check-indices | time="2024-01-22T06:54:33Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 06:54:43 | es-rollover-autoprov/11-check-indices | time="2024-01-22T06:54:43Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 06:54:53 | es-rollover-autoprov/11-check-indices | time="2024-01-22T06:54:53Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 06:55:03 | es-rollover-autoprov/11-check-indices | time="2024-01-22T06:55:03Z" level=info msg="Job of owner my-jaeger-es-rollover succeeded after my-jaeger-es-rollover 40.016530835s" logger.go:42: 06:55:04 | es-rollover-autoprov/11-check-indices | Job:kuttl-test-busy-turkey/05-check-indices created logger.go:42: 06:55:08 | es-rollover-autoprov/11-check-indices | test step completed 11-check-indices logger.go:42: 06:55:08 | es-rollover-autoprov/12-check-indices | starting test step 12-check-indices logger.go:42: 06:55:08 | es-rollover-autoprov/12-check-indices | Job:kuttl-test-busy-turkey/06-check-indices created logger.go:42: 06:55:12 | es-rollover-autoprov/12-check-indices | test step completed 12-check-indices logger.go:42: 06:55:12 | es-rollover-autoprov | es-rollover-autoprov events from ns kuttl-test-busy-turkey: logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:27 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestbusyturkeymyjaeger-1-5bc665b885 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestbusyturkeymyjaeger-1-5bc665b885v9nv9 replicaset-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusyturkeymyjaeger-1-5bc665b885v9nv9 Binding Scheduled Successfully assigned kuttl-test-busy-turkey/elasticsearch-cdm-kuttltestbusyturkeymyjaeger-1-5bc665b885v9nv9 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:27 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestbusyturkeymyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestbusyturkeymyjaeger-1-5bc665b885 to 1 deployment-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusyturkeymyjaeger-1-5bc665b885v9nv9 AddedInterface Add eth0 [10.131.0.19/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusyturkeymyjaeger-1-5bc665b885v9nv9.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:3ad088cbfcfb006d87762e96a70ff989e75d49872954347f7efe34d9529c1b7a" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusyturkeymyjaeger-1-5bc665b885v9nv9.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusyturkeymyjaeger-1-5bc665b885v9nv9.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusyturkeymyjaeger-1-5bc665b885v9nv9.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:911ee3f26ae71eab71bf2a1455d1c2ea871cddc78a51189e3ad2f18f63bd3d91" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusyturkeymyjaeger-1-5bc665b885v9nv9.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestbusyturkeymyjaeger-1-5bc665b885v9nv9.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:43 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestbusyturkeymyjaeger-1-5bc665b885v9nv9.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:54 +0000 UTC Normal Pod my-jaeger-collector-7686dd68ff-ps726 Binding Scheduled Successfully assigned kuttl-test-busy-turkey/my-jaeger-collector-7686dd68ff-ps726 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:54 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7686dd68ff SuccessfulCreate Created pod: my-jaeger-collector-7686dd68ff-ps726 replicaset-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:54 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7686dd68ff to 1 deployment-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:54 +0000 UTC Normal Pod my-jaeger-query-7899fb97-sk4nl Binding Scheduled Successfully assigned kuttl-test-busy-turkey/my-jaeger-query-7899fb97-sk4nl to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:54 +0000 UTC Normal Pod my-jaeger-query-7899fb97-sk4nl AddedInterface Add eth0 [10.129.2.19/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:54 +0000 UTC Normal Pod my-jaeger-query-7899fb97-sk4nl.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:54 +0000 UTC Normal Pod my-jaeger-query-7899fb97-sk4nl.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:54 +0000 UTC Normal Pod my-jaeger-query-7899fb97-sk4nl.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:54 +0000 UTC Normal Pod my-jaeger-query-7899fb97-sk4nl.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:54 +0000 UTC Normal Pod my-jaeger-query-7899fb97-sk4nl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:54 +0000 UTC Normal Pod my-jaeger-query-7899fb97-sk4nl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:54 +0000 UTC Normal Pod my-jaeger-query-7899fb97-sk4nl.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:54 +0000 UTC Normal Pod my-jaeger-query-7899fb97-sk4nl.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:54 +0000 UTC Normal Pod my-jaeger-query-7899fb97-sk4nl.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:54 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7899fb97 SuccessfulCreate Created pod: my-jaeger-query-7899fb97-sk4nl replicaset-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:54 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7899fb97 to 1 deployment-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:55 +0000 UTC Normal Pod my-jaeger-collector-7686dd68ff-ps726 AddedInterface Add eth0 [10.128.2.30/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:55 +0000 UTC Normal Pod my-jaeger-collector-7686dd68ff-ps726.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:55 +0000 UTC Normal Pod my-jaeger-collector-7686dd68ff-ps726.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:51:55 +0000 UTC Normal Pod my-jaeger-collector-7686dd68ff-ps726.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:01 +0000 UTC Normal Pod my-jaeger-query-7899fb97-sk4nl.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:01 +0000 UTC Normal Pod my-jaeger-query-7899fb97-sk4nl.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:01 +0000 UTC Normal Pod my-jaeger-query-7899fb97-sk4nl.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:01 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7899fb97 SuccessfulDelete Deleted pod: my-jaeger-query-7899fb97-sk4nl replicaset-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:01 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-7899fb97 to 0 from 1 deployment-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:02 +0000 UTC Normal Pod my-jaeger-query-788c84497f-vdshb Binding Scheduled Successfully assigned kuttl-test-busy-turkey/my-jaeger-query-788c84497f-vdshb to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:02 +0000 UTC Normal Pod my-jaeger-query-788c84497f-vdshb AddedInterface Add eth0 [10.129.2.20/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:02 +0000 UTC Normal Pod my-jaeger-query-788c84497f-vdshb.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:02 +0000 UTC Normal Pod my-jaeger-query-788c84497f-vdshb.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:02 +0000 UTC Normal Pod my-jaeger-query-788c84497f-vdshb.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:02 +0000 UTC Normal Pod my-jaeger-query-788c84497f-vdshb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:02 +0000 UTC Normal Pod my-jaeger-query-788c84497f-vdshb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:02 +0000 UTC Normal Pod my-jaeger-query-788c84497f-vdshb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:02 +0000 UTC Normal Pod my-jaeger-query-788c84497f-vdshb.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:02 +0000 UTC Normal Pod my-jaeger-query-788c84497f-vdshb.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:02 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-788c84497f SuccessfulCreate Created pod: my-jaeger-query-788c84497f-vdshb replicaset-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:02 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-788c84497f to 1 deployment-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:03 +0000 UTC Normal Pod my-jaeger-query-788c84497f-vdshb.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:07 +0000 UTC Normal Pod 00-report-span-bjpxl Binding Scheduled Successfully assigned kuttl-test-busy-turkey/00-report-span-bjpxl to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:07 +0000 UTC Normal Pod 00-report-span-bjpxl AddedInterface Add eth0 [10.128.2.31/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:07 +0000 UTC Normal Pod 00-report-span-bjpxl.spec.containers{asserts-container} Pulling Pulling image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:07 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-bjpxl job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:09 +0000 UTC Normal Pod 00-report-span-bjpxl.spec.containers{asserts-container} Pulled Successfully pulled image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" in 1.615s (1.615s including waiting) kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:09 +0000 UTC Normal Pod 00-report-span-bjpxl.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:09 +0000 UTC Normal Pod 00-report-span-bjpxl.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:09 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:09 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:09 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:32 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:33 +0000 UTC Normal Pod 00-check-indices-69cdj Binding Scheduled Successfully assigned kuttl-test-busy-turkey/00-check-indices-69cdj to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:33 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-69cdj job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:34 +0000 UTC Normal Pod 00-check-indices-69cdj AddedInterface Add eth0 [10.128.2.32/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:34 +0000 UTC Normal Pod 00-check-indices-69cdj.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:34 +0000 UTC Normal Pod 00-check-indices-69cdj.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:34 +0000 UTC Normal Pod 00-check-indices-69cdj.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:36 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:37 +0000 UTC Normal Pod 01-check-indices-vdjv8 Binding Scheduled Successfully assigned kuttl-test-busy-turkey/01-check-indices-vdjv8 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:37 +0000 UTC Normal Job.batch 01-check-indices SuccessfulCreate Created pod: 01-check-indices-vdjv8 job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:38 +0000 UTC Normal Pod 01-check-indices-vdjv8 AddedInterface Add eth0 [10.128.2.33/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:38 +0000 UTC Normal Pod 01-check-indices-vdjv8.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:38 +0000 UTC Normal Pod 01-check-indices-vdjv8.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:38 +0000 UTC Normal Pod 01-check-indices-vdjv8.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:39 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:39 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-7686dd68ff-ps726 horizontal-pod-autoscaler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:39 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:40 +0000 UTC Normal Job.batch 01-check-indices Completed Job completed job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:42 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-v7z4w Binding Scheduled Successfully assigned kuttl-test-busy-turkey/my-jaeger-es-rollover-create-mapping-v7z4w to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:42 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping SuccessfulCreate Created pod: my-jaeger-es-rollover-create-mapping-v7z4w job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:43 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-v7z4w AddedInterface Add eth0 [10.128.2.34/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:43 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-v7z4w.spec.containers{my-jaeger-es-rollover-create-mapping} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:47 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-v7z4w.spec.containers{my-jaeger-es-rollover-create-mapping} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" in 4.673s (4.673s including waiting) kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:48 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-v7z4w.spec.containers{my-jaeger-es-rollover-create-mapping} Created Created container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:48 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-v7z4w.spec.containers{my-jaeger-es-rollover-create-mapping} Started Started container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:50 +0000 UTC Normal Pod my-jaeger-collector-7686dd68ff-ps726.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:50 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7686dd68ff SuccessfulDelete Deleted pod: my-jaeger-collector-7686dd68ff-ps726 replicaset-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:50 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-7686dd68ff to 0 from 1 deployment-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:50 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping Completed Job completed job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:50 +0000 UTC Normal Pod my-jaeger-query-788c84497f-vdshb.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:50 +0000 UTC Normal Pod my-jaeger-query-788c84497f-vdshb.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:50 +0000 UTC Normal Pod my-jaeger-query-788c84497f-vdshb.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:50 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-788c84497f SuccessfulDelete Deleted pod: my-jaeger-query-788c84497f-vdshb replicaset-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:50 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-788c84497f to 0 from 1 deployment-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:51 +0000 UTC Normal Pod my-jaeger-collector-65bd5bbcb-c9gv9 Binding Scheduled Successfully assigned kuttl-test-busy-turkey/my-jaeger-collector-65bd5bbcb-c9gv9 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:51 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-65bd5bbcb SuccessfulCreate Created pod: my-jaeger-collector-65bd5bbcb-c9gv9 replicaset-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:51 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-65bd5bbcb to 1 deployment-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:51 +0000 UTC Normal Pod my-jaeger-query-57488874d6-cmtsl Binding Scheduled Successfully assigned kuttl-test-busy-turkey/my-jaeger-query-57488874d6-cmtsl to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:51 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-57488874d6 SuccessfulCreate Created pod: my-jaeger-query-57488874d6-cmtsl replicaset-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:51 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-57488874d6 to 1 deployment-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:52 +0000 UTC Normal Pod my-jaeger-collector-65bd5bbcb-c9gv9 AddedInterface Add eth0 [10.128.2.35/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:52 +0000 UTC Normal Pod my-jaeger-collector-65bd5bbcb-c9gv9.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:52 +0000 UTC Normal Pod my-jaeger-collector-65bd5bbcb-c9gv9.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:52 +0000 UTC Normal Pod my-jaeger-collector-65bd5bbcb-c9gv9.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:52 +0000 UTC Normal Pod my-jaeger-query-57488874d6-cmtsl AddedInterface Add eth0 [10.129.2.21/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:52 +0000 UTC Normal Pod my-jaeger-query-57488874d6-cmtsl.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:52 +0000 UTC Normal Pod my-jaeger-query-57488874d6-cmtsl.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:52 +0000 UTC Normal Pod my-jaeger-query-57488874d6-cmtsl.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:52 +0000 UTC Normal Pod my-jaeger-query-57488874d6-cmtsl.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:52 +0000 UTC Normal Pod my-jaeger-query-57488874d6-cmtsl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:52 +0000 UTC Normal Pod my-jaeger-query-57488874d6-cmtsl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:52 +0000 UTC Normal Pod my-jaeger-query-57488874d6-cmtsl.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:52 +0000 UTC Normal Pod my-jaeger-query-57488874d6-cmtsl.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:52:52 +0000 UTC Normal Pod my-jaeger-query-57488874d6-cmtsl.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:00 +0000 UTC Normal Pod 02-report-span-grxgt Binding Scheduled Successfully assigned kuttl-test-busy-turkey/02-report-span-grxgt to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:00 +0000 UTC Normal Job.batch 02-report-span SuccessfulCreate Created pod: 02-report-span-grxgt job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28431773-vtzpg Binding Scheduled Successfully assigned kuttl-test-busy-turkey/my-jaeger-es-lookback-28431773-vtzpg to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28431773-vtzpg AddedInterface Add eth0 [10.128.2.36/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28431773-vtzpg.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28431773-vtzpg.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28431773-vtzpg.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28431773 SuccessfulCreate Created pod: my-jaeger-es-lookback-28431773-vtzpg job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28431773 cronjob-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28431773-kl5ht Binding Scheduled Successfully assigned kuttl-test-busy-turkey/my-jaeger-es-rollover-28431773-kl5ht to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28431773-kl5ht AddedInterface Add eth0 [10.128.2.37/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28431773-kl5ht.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28431773-kl5ht.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28431773-kl5ht.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28431773 SuccessfulCreate Created pod: my-jaeger-es-rollover-28431773-kl5ht job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28431773 cronjob-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:01 +0000 UTC Normal Pod 02-report-span-grxgt AddedInterface Add eth0 [10.128.2.38/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:01 +0000 UTC Normal Pod 02-report-span-grxgt.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:01 +0000 UTC Normal Pod 02-report-span-grxgt.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:01 +0000 UTC Normal Pod 02-report-span-grxgt.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28431773 Completed Job completed job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28431773, status: Complete cronjob-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28431773 Completed Job completed job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28431773, status: Complete cronjob-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:24 +0000 UTC Normal Job.batch 02-report-span Completed Job completed job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:25 +0000 UTC Normal Pod 02-check-indices-cl5fz Binding Scheduled Successfully assigned kuttl-test-busy-turkey/02-check-indices-cl5fz to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:25 +0000 UTC Normal Pod 02-check-indices-cl5fz AddedInterface Add eth0 [10.128.2.39/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:25 +0000 UTC Normal Pod 02-check-indices-cl5fz.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:25 +0000 UTC Normal Pod 02-check-indices-cl5fz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:25 +0000 UTC Normal Pod 02-check-indices-cl5fz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:25 +0000 UTC Normal Job.batch 02-check-indices SuccessfulCreate Created pod: 02-check-indices-cl5fz job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:28 +0000 UTC Normal Job.batch 02-check-indices Completed Job completed job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:29 +0000 UTC Normal Pod 03-check-indices-nkzsd Binding Scheduled Successfully assigned kuttl-test-busy-turkey/03-check-indices-nkzsd to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:29 +0000 UTC Normal Pod 03-check-indices-nkzsd AddedInterface Add eth0 [10.128.2.40/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:29 +0000 UTC Normal Pod 03-check-indices-nkzsd.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:29 +0000 UTC Normal Pod 03-check-indices-nkzsd.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:29 +0000 UTC Normal Pod 03-check-indices-nkzsd.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:29 +0000 UTC Normal Job.batch 03-check-indices SuccessfulCreate Created pod: 03-check-indices-nkzsd job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:32 +0000 UTC Normal Job.batch 03-check-indices Completed Job completed job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:33 +0000 UTC Normal Pod 04-check-indices-6mjr8 Binding Scheduled Successfully assigned kuttl-test-busy-turkey/04-check-indices-6mjr8 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:33 +0000 UTC Normal Pod 04-check-indices-6mjr8 AddedInterface Add eth0 [10.128.2.41/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:33 +0000 UTC Normal Pod 04-check-indices-6mjr8.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:33 +0000 UTC Normal Job.batch 04-check-indices SuccessfulCreate Created pod: 04-check-indices-6mjr8 job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:34 +0000 UTC Normal Pod 04-check-indices-6mjr8.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:34 +0000 UTC Normal Pod 04-check-indices-6mjr8.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:36 +0000 UTC Normal Job.batch 04-check-indices Completed Job completed job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:39 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-65bd5bbcb-c9gv9 horizontal-pod-autoscaler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:47 +0000 UTC Normal Pod 03-report-span-nj2lv Binding Scheduled Successfully assigned kuttl-test-busy-turkey/03-report-span-nj2lv to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:47 +0000 UTC Normal Job.batch 03-report-span SuccessfulCreate Created pod: 03-report-span-nj2lv job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:48 +0000 UTC Normal Pod 03-report-span-nj2lv AddedInterface Add eth0 [10.128.2.42/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:48 +0000 UTC Normal Pod 03-report-span-nj2lv.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:48 +0000 UTC Normal Pod 03-report-span-nj2lv.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:53:48 +0000 UTC Normal Pod 03-report-span-nj2lv.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28431774-22gld Binding Scheduled Successfully assigned kuttl-test-busy-turkey/my-jaeger-es-lookback-28431774-22gld to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28431774-22gld AddedInterface Add eth0 [10.128.2.44/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28431774-22gld.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28431774-22gld.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28431774-22gld.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28431774 SuccessfulCreate Created pod: my-jaeger-es-lookback-28431774-22gld job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28431774 cronjob-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28431774-w7q8t Binding Scheduled Successfully assigned kuttl-test-busy-turkey/my-jaeger-es-rollover-28431774-w7q8t to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28431774-w7q8t AddedInterface Add eth0 [10.128.2.43/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28431774-w7q8t.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28431774-w7q8t.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28431774-w7q8t.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28431774 SuccessfulCreate Created pod: my-jaeger-es-rollover-28431774-w7q8t job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28431774 cronjob-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28431774 Completed Job completed job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28431774, status: Complete cronjob-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:04 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28431774 Completed Job completed job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:04 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28431774, status: Complete cronjob-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:54:10 +0000 UTC Normal Job.batch 03-report-span Completed Job completed job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28431775-nkk92 Binding Scheduled Successfully assigned kuttl-test-busy-turkey/my-jaeger-es-lookback-28431775-nkk92 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28431775-nkk92 AddedInterface Add eth0 [10.128.2.46/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28431775-nkk92.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28431775-nkk92.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28431775-nkk92.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28431775 SuccessfulCreate Created pod: my-jaeger-es-lookback-28431775-nkk92 job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28431775 cronjob-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28431775-wdlhl Binding Scheduled Successfully assigned kuttl-test-busy-turkey/my-jaeger-es-rollover-28431775-wdlhl to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28431775-wdlhl AddedInterface Add eth0 [10.128.2.45/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28431775-wdlhl.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:011593d9ab270edcf57a484b7106c1682984a35131a92deabcb944aefb180619" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28431775-wdlhl.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28431775-wdlhl.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28431775 SuccessfulCreate Created pod: my-jaeger-es-rollover-28431775-wdlhl job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28431775 cronjob-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28431775 Completed Job completed job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28431775, status: Complete cronjob-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:04 +0000 UTC Normal Pod 05-check-indices-455cp Binding Scheduled Successfully assigned kuttl-test-busy-turkey/05-check-indices-455cp to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:04 +0000 UTC Normal Pod 05-check-indices-455cp AddedInterface Add eth0 [10.128.2.47/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:04 +0000 UTC Normal Pod 05-check-indices-455cp.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:04 +0000 UTC Normal Pod 05-check-indices-455cp.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:04 +0000 UTC Normal Pod 05-check-indices-455cp.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:04 +0000 UTC Normal Job.batch 05-check-indices SuccessfulCreate Created pod: 05-check-indices-455cp job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:04 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28431775 Completed Job completed job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:04 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28431775, status: Complete cronjob-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:07 +0000 UTC Normal Job.batch 05-check-indices Completed Job completed job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:08 +0000 UTC Normal Pod 06-check-indices-7824b Binding Scheduled Successfully assigned kuttl-test-busy-turkey/06-check-indices-7824b to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:08 +0000 UTC Normal Pod 06-check-indices-7824b AddedInterface Add eth0 [10.128.2.48/23] from ovn-kubernetes logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:08 +0000 UTC Normal Pod 06-check-indices-7824b.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:08 +0000 UTC Normal Pod 06-check-indices-7824b.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:08 +0000 UTC Normal Pod 06-check-indices-7824b.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:08 +0000 UTC Normal Job.batch 06-check-indices SuccessfulCreate Created pod: 06-check-indices-7824b job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | 2024-01-22 06:55:11 +0000 UTC Normal Job.batch 06-check-indices Completed Job completed job-controller logger.go:42: 06:55:12 | es-rollover-autoprov | Deleting namespace: kuttl-test-busy-turkey === CONT kuttl/harness/es-increasing-replicas logger.go:42: 06:55:19 | es-increasing-replicas | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 06:55:19 | es-increasing-replicas | Ignoring check-es-nodes.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 06:55:19 | es-increasing-replicas | Ignoring openshift-check-es-nodes.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 06:55:19 | es-increasing-replicas | Creating namespace: kuttl-test-guided-narwhal logger.go:42: 06:55:19 | es-increasing-replicas/1-install | starting test step 1-install logger.go:42: 06:55:19 | es-increasing-replicas/1-install | Jaeger:kuttl-test-guided-narwhal/simple-prod created logger.go:42: 06:55:58 | es-increasing-replicas/1-install | test step completed 1-install logger.go:42: 06:55:58 | es-increasing-replicas/2-install | starting test step 2-install logger.go:42: 06:55:58 | es-increasing-replicas/2-install | Jaeger:kuttl-test-guided-narwhal/simple-prod updated logger.go:42: 06:56:18 | es-increasing-replicas/2-install | test step completed 2-install logger.go:42: 06:56:18 | es-increasing-replicas/3-smoke-test | starting test step 3-smoke-test logger.go:42: 06:56:18 | es-increasing-replicas/3-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 06:56:20 | es-increasing-replicas/3-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 06:56:26 | es-increasing-replicas/3-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 06:56:27 | es-increasing-replicas/3-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 06:56:27 | es-increasing-replicas/3-smoke-test | job.batch/report-span created logger.go:42: 06:56:27 | es-increasing-replicas/3-smoke-test | job.batch/check-span created logger.go:42: 06:56:38 | es-increasing-replicas/3-smoke-test | test step completed 3-smoke-test logger.go:42: 06:56:38 | es-increasing-replicas/4-install | starting test step 4-install logger.go:42: 06:56:38 | es-increasing-replicas/4-install | Jaeger:kuttl-test-guided-narwhal/simple-prod updated logger.go:42: 06:56:38 | es-increasing-replicas/4-install | test step completed 4-install logger.go:42: 06:56:38 | es-increasing-replicas/5-check-es-nodes | starting test step 5-check-es-nodes logger.go:42: 06:56:38 | es-increasing-replicas/5-check-es-nodes | running command: [sh -c ./check-es-nodes.sh $NAMESPACE] logger.go:42: 06:56:38 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 06:56:38 | es-increasing-replicas/5-check-es-nodes | false logger.go:42: 06:56:38 | es-increasing-replicas/5-check-es-nodes | Error: no matches found logger.go:42: 06:56:43 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 06:56:43 | es-increasing-replicas/5-check-es-nodes | true logger.go:42: 06:56:43 | es-increasing-replicas/5-check-es-nodes | test step completed 5-check-es-nodes logger.go:42: 06:56:43 | es-increasing-replicas | es-increasing-replicas events from ns kuttl-test-guided-narwhal: logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:26 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-1-7bb8c9468c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-1-7bb8cxtr8h replicaset-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-1-7bb8cxtr8h Binding Scheduled Successfully assigned kuttl-test-guided-narwhal/elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-1-7bb8cxtr8h to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:26 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-1-7bb8cxtr8h FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:26 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-1-7bb8c9468c to 1 deployment-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-1-7bb8cxtr8h AddedInterface Add eth0 [10.131.0.20/23] from ovn-kubernetes logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-1-7bb8cxtr8h.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:3ad088cbfcfb006d87762e96a70ff989e75d49872954347f7efe34d9529c1b7a" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-1-7bb8cxtr8h.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-1-7bb8cxtr8h.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-1-7bb8cxtr8h.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:911ee3f26ae71eab71bf2a1455d1c2ea871cddc78a51189e3ad2f18f63bd3d91" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-1-7bb8cxtr8h.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-1-7bb8cxtr8h.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:37 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-1-7bb8cxtr8h.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:42 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-1-7bb8cxtr8h.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:53 +0000 UTC Normal Pod simple-prod-collector-5c78555b4f-wj5b4 Binding Scheduled Successfully assigned kuttl-test-guided-narwhal/simple-prod-collector-5c78555b4f-wj5b4 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:53 +0000 UTC Warning Pod simple-prod-collector-5c78555b4f-wj5b4 FailedMount MountVolume.SetUp failed for volume "simple-prod-collector-tls-config-volume" : secret "simple-prod-collector-headless-tls" not found kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:53 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5c78555b4f SuccessfulCreate Created pod: simple-prod-collector-5c78555b4f-wj5b4 replicaset-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:53 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-5c78555b4f to 1 deployment-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:53 +0000 UTC Normal Pod simple-prod-query-5699959d75-fkqw7 Binding Scheduled Successfully assigned kuttl-test-guided-narwhal/simple-prod-query-5699959d75-fkqw7 to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:53 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5699959d75 SuccessfulCreate Created pod: simple-prod-query-5699959d75-fkqw7 replicaset-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:53 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-5699959d75 to 1 deployment-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:54 +0000 UTC Normal Pod simple-prod-collector-5c78555b4f-wj5b4 AddedInterface Add eth0 [10.128.2.49/23] from ovn-kubernetes logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:54 +0000 UTC Normal Pod simple-prod-collector-5c78555b4f-wj5b4.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:54 +0000 UTC Normal Pod simple-prod-collector-5c78555b4f-wj5b4.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:54 +0000 UTC Normal Pod simple-prod-collector-5c78555b4f-wj5b4.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:54 +0000 UTC Warning Pod simple-prod-query-5699959d75-fkqw7 FailedMount MountVolume.SetUp failed for volume "simple-prod-ui-oauth-proxy-tls" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:55 +0000 UTC Normal Pod simple-prod-query-5699959d75-fkqw7 AddedInterface Add eth0 [10.129.2.22/23] from ovn-kubernetes logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:55 +0000 UTC Normal Pod simple-prod-query-5699959d75-fkqw7.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:55 +0000 UTC Normal Pod simple-prod-query-5699959d75-fkqw7.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:55 +0000 UTC Normal Pod simple-prod-query-5699959d75-fkqw7.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:55 +0000 UTC Normal Pod simple-prod-query-5699959d75-fkqw7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:55 +0000 UTC Normal Pod simple-prod-query-5699959d75-fkqw7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:55 +0000 UTC Normal Pod simple-prod-query-5699959d75-fkqw7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:55 +0000 UTC Normal Pod simple-prod-query-5699959d75-fkqw7.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:56 +0000 UTC Normal Pod simple-prod-query-5699959d75-fkqw7.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:56 +0000 UTC Normal Pod simple-prod-query-5699959d75-fkqw7.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:59 +0000 UTC Normal Pod simple-prod-collector-5c78555b4f-wkbrs Binding Scheduled Successfully assigned kuttl-test-guided-narwhal/simple-prod-collector-5c78555b4f-wkbrs to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:59 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5c78555b4f SuccessfulCreate Created pod: simple-prod-collector-5c78555b4f-wkbrs replicaset-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:59 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-5c78555b4f to 2 from 1 deployment-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:59 +0000 UTC Normal Pod simple-prod-query-5699959d75-5fq6g Binding Scheduled Successfully assigned kuttl-test-guided-narwhal/simple-prod-query-5699959d75-5fq6g to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:59 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5699959d75 SuccessfulCreate Created pod: simple-prod-query-5699959d75-5fq6g replicaset-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:55:59 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-5699959d75 to 2 from 1 deployment-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:00 +0000 UTC Normal Pod simple-prod-collector-5c78555b4f-wkbrs AddedInterface Add eth0 [10.131.0.21/23] from ovn-kubernetes logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:00 +0000 UTC Normal Pod simple-prod-collector-5c78555b4f-wkbrs.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:00 +0000 UTC Normal Pod simple-prod-query-5699959d75-5fq6g AddedInterface Add eth0 [10.131.0.22/23] from ovn-kubernetes logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:00 +0000 UTC Normal Pod simple-prod-query-5699959d75-5fq6g.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:07 +0000 UTC Normal Pod simple-prod-collector-5c78555b4f-wkbrs.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" in 7.102s (7.102s including waiting) kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:07 +0000 UTC Normal Pod simple-prod-collector-5c78555b4f-wkbrs.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:07 +0000 UTC Normal Pod simple-prod-collector-5c78555b4f-wkbrs.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:07 +0000 UTC Normal Pod simple-prod-query-5699959d75-5fq6g.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" in 6.804s (6.804s including waiting) kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:07 +0000 UTC Normal Pod simple-prod-query-5699959d75-5fq6g.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:07 +0000 UTC Normal Pod simple-prod-query-5699959d75-5fq6g.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:07 +0000 UTC Normal Pod simple-prod-query-5699959d75-5fq6g.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:12 +0000 UTC Normal Pod simple-prod-query-5699959d75-5fq6g.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" in 5.576s (5.576s including waiting) kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:12 +0000 UTC Normal Pod simple-prod-query-5699959d75-5fq6g.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:13 +0000 UTC Normal Pod simple-prod-query-5699959d75-5fq6g.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:13 +0000 UTC Normal Pod simple-prod-query-5699959d75-5fq6g.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:15 +0000 UTC Normal Pod simple-prod-query-5699959d75-5fq6g.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" in 2.849s (2.849s including waiting) kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:15 +0000 UTC Normal Pod simple-prod-query-5699959d75-5fq6g.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:15 +0000 UTC Normal Pod simple-prod-query-5699959d75-5fq6g.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:21 +0000 UTC Normal Pod simple-prod-query-5699959d75-5fq6g.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:21 +0000 UTC Normal Pod simple-prod-query-5699959d75-5fq6g.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:21 +0000 UTC Normal Pod simple-prod-query-5699959d75-5fq6g.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:21 +0000 UTC Normal Pod simple-prod-query-5699959d75-fkqw7.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:21 +0000 UTC Normal Pod simple-prod-query-5699959d75-fkqw7.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:21 +0000 UTC Normal Pod simple-prod-query-5699959d75-fkqw7.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:21 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5699959d75 SuccessfulDelete Deleted pod: simple-prod-query-5699959d75-fkqw7 replicaset-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:21 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5699959d75 SuccessfulDelete Deleted pod: simple-prod-query-5699959d75-5fq6g replicaset-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:21 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-5699959d75 to 0 from 2 deployment-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:22 +0000 UTC Normal Pod simple-prod-query-7879b4b955-htpql Binding Scheduled Successfully assigned kuttl-test-guided-narwhal/simple-prod-query-7879b4b955-htpql to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:22 +0000 UTC Normal Pod simple-prod-query-7879b4b955-zpgpl Binding Scheduled Successfully assigned kuttl-test-guided-narwhal/simple-prod-query-7879b4b955-zpgpl to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:22 +0000 UTC Normal ReplicaSet.apps simple-prod-query-7879b4b955 SuccessfulCreate Created pod: simple-prod-query-7879b4b955-htpql replicaset-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:22 +0000 UTC Normal ReplicaSet.apps simple-prod-query-7879b4b955 SuccessfulCreate Created pod: simple-prod-query-7879b4b955-zpgpl replicaset-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:22 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-7879b4b955 to 2 deployment-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-htpql AddedInterface Add eth0 [10.129.2.23/23] from ovn-kubernetes logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-htpql.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-htpql.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-htpql.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-htpql.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-htpql.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-htpql.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-htpql.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-htpql.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-htpql.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-zpgpl AddedInterface Add eth0 [10.131.0.23/23] from ovn-kubernetes logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-zpgpl.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-zpgpl.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-zpgpl.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-zpgpl.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-zpgpl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-zpgpl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-zpgpl.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-zpgpl.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:23 +0000 UTC Normal Pod simple-prod-query-7879b4b955-zpgpl.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:27 +0000 UTC Normal Pod check-span-jt2mf Binding Scheduled Successfully assigned kuttl-test-guided-narwhal/check-span-jt2mf to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:27 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-jt2mf job-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:27 +0000 UTC Normal Pod report-span-v8hrn Binding Scheduled Successfully assigned kuttl-test-guided-narwhal/report-span-v8hrn to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:27 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-v8hrn job-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:28 +0000 UTC Normal Pod check-span-jt2mf AddedInterface Add eth0 [10.128.2.51/23] from ovn-kubernetes logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:28 +0000 UTC Normal Pod check-span-jt2mf.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:28 +0000 UTC Normal Pod check-span-jt2mf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:28 +0000 UTC Normal Pod check-span-jt2mf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:28 +0000 UTC Normal Pod report-span-v8hrn AddedInterface Add eth0 [10.128.2.50/23] from ovn-kubernetes logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:28 +0000 UTC Normal Pod report-span-v8hrn.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:28 +0000 UTC Normal Pod report-span-v8hrn.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:28 +0000 UTC Normal Pod report-span-v8hrn.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:38 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:39 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-2-cf8f76f79 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-2-cf8f78x4d9 replicaset-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-2-cf8f78x4d9 Binding Scheduled Successfully assigned kuttl-test-guided-narwhal/elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-2-cf8f78x4d9 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:39 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-2 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-2-cf8f76f79 to 1 deployment-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:39 +0000 UTC Normal Pod simple-prod-collector-5c78555b4f-wj5b4.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:39 +0000 UTC Normal Pod simple-prod-collector-5c78555b4f-wkbrs.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:39 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5c78555b4f SuccessfulDelete Deleted pod: simple-prod-collector-5c78555b4f-wj5b4 replicaset-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:39 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5c78555b4f SuccessfulDelete Deleted pod: simple-prod-collector-5c78555b4f-wkbrs replicaset-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:39 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled down replica set simple-prod-collector-5c78555b4f to 0 from 2 deployment-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:39 +0000 UTC Normal Pod simple-prod-query-7879b4b955-htpql.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:39 +0000 UTC Normal Pod simple-prod-query-7879b4b955-htpql.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:39 +0000 UTC Normal Pod simple-prod-query-7879b4b955-htpql.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:39 +0000 UTC Normal Pod simple-prod-query-7879b4b955-zpgpl.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:39 +0000 UTC Normal Pod simple-prod-query-7879b4b955-zpgpl.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:39 +0000 UTC Normal Pod simple-prod-query-7879b4b955-zpgpl.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:39 +0000 UTC Normal ReplicaSet.apps simple-prod-query-7879b4b955 SuccessfulDelete Deleted pod: simple-prod-query-7879b4b955-htpql replicaset-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:39 +0000 UTC Normal ReplicaSet.apps simple-prod-query-7879b4b955 SuccessfulDelete Deleted pod: simple-prod-query-7879b4b955-zpgpl replicaset-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:39 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-7879b4b955 to 0 from 2 deployment-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-2-cf8f78x4d9 AddedInterface Add eth0 [10.128.2.52/23] from ovn-kubernetes logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-2-cf8f78x4d9.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:3ad088cbfcfb006d87762e96a70ff989e75d49872954347f7efe34d9529c1b7a" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-2-cf8f78x4d9.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-2-cf8f78x4d9.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-2-cf8f78x4d9.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:911ee3f26ae71eab71bf2a1455d1c2ea871cddc78a51189e3ad2f18f63bd3d91" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-2-cf8f78x4d9.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestguidednarwhalsimpleprod-2-cf8f78x4d9.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:40 +0000 UTC Normal Pod simple-prod-collector-79f75b9cc8-c75vv Binding Scheduled Successfully assigned kuttl-test-guided-narwhal/simple-prod-collector-79f75b9cc8-c75vv to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:40 +0000 UTC Normal Pod simple-prod-collector-79f75b9cc8-rnrns Binding Scheduled Successfully assigned kuttl-test-guided-narwhal/simple-prod-collector-79f75b9cc8-rnrns to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:40 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-79f75b9cc8 SuccessfulCreate Created pod: simple-prod-collector-79f75b9cc8-rnrns replicaset-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:40 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-79f75b9cc8 SuccessfulCreate Created pod: simple-prod-collector-79f75b9cc8-c75vv replicaset-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:40 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-79f75b9cc8 to 2 deployment-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:40 +0000 UTC Normal Pod simple-prod-query-65c68559f8-fhdv6 Binding Scheduled Successfully assigned kuttl-test-guided-narwhal/simple-prod-query-65c68559f8-fhdv6 to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:40 +0000 UTC Normal Pod simple-prod-query-65c68559f8-qrvx2 Binding Scheduled Successfully assigned kuttl-test-guided-narwhal/simple-prod-query-65c68559f8-qrvx2 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:40 +0000 UTC Normal ReplicaSet.apps simple-prod-query-65c68559f8 SuccessfulCreate Created pod: simple-prod-query-65c68559f8-fhdv6 replicaset-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:40 +0000 UTC Normal ReplicaSet.apps simple-prod-query-65c68559f8 SuccessfulCreate Created pod: simple-prod-query-65c68559f8-qrvx2 replicaset-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:40 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-65c68559f8 to 2 deployment-controller logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-collector-79f75b9cc8-c75vv AddedInterface Add eth0 [10.131.0.24/23] from ovn-kubernetes logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-collector-79f75b9cc8-c75vv.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-collector-79f75b9cc8-c75vv.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-collector-79f75b9cc8-c75vv.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-collector-79f75b9cc8-rnrns AddedInterface Add eth0 [10.129.2.24/23] from ovn-kubernetes logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-collector-79f75b9cc8-rnrns.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-collector-79f75b9cc8-rnrns.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-collector-79f75b9cc8-rnrns.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-fhdv6 AddedInterface Add eth0 [10.129.2.25/23] from ovn-kubernetes logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-fhdv6.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-fhdv6.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-fhdv6.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-fhdv6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-fhdv6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-fhdv6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-fhdv6.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-fhdv6.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-fhdv6.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-qrvx2 AddedInterface Add eth0 [10.131.0.25/23] from ovn-kubernetes logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-qrvx2.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-qrvx2.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-qrvx2.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-qrvx2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-qrvx2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-qrvx2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-qrvx2.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-qrvx2.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:41 +0000 UTC Normal Pod simple-prod-query-65c68559f8-qrvx2.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 06:56:43 | es-increasing-replicas | 2024-01-22 06:56:43 +0000 UTC Warning Endpoints simple-prod-collector FailedToUpdateEndpoint Failed to update endpoint kuttl-test-guided-narwhal/simple-prod-collector: Operation cannot be fulfilled on endpoints "simple-prod-collector": the object has been modified; please apply your changes to the latest version and try again endpoint-controller logger.go:42: 06:56:43 | es-increasing-replicas | Deleting namespace: kuttl-test-guided-narwhal === CONT kuttl/harness/es-index-cleaner-autoprov logger.go:42: 06:57:19 | es-index-cleaner-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 06:57:19 | es-index-cleaner-autoprov | Creating namespace: kuttl-test-internal-wombat logger.go:42: 06:57:19 | es-index-cleaner-autoprov/1-install | starting test step 1-install logger.go:42: 06:57:19 | es-index-cleaner-autoprov/1-install | Jaeger:kuttl-test-internal-wombat/test-es-index-cleaner-with-prefix created logger.go:42: 06:57:56 | es-index-cleaner-autoprov/1-install | test step completed 1-install logger.go:42: 06:57:56 | es-index-cleaner-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 06:57:56 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE test-es-index-cleaner-with-prefix /dev/null] logger.go:42: 06:57:57 | es-index-cleaner-autoprov/2-report-spans | Warning: resource jaegers/test-es-index-cleaner-with-prefix is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 06:58:03 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c DAYS=5 ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 06:58:04 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 06:58:04 | es-index-cleaner-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 06:58:43 | es-index-cleaner-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 06:58:43 | es-index-cleaner-autoprov/3-install | starting test step 3-install logger.go:42: 06:58:43 | es-index-cleaner-autoprov/3-install | Jaeger:kuttl-test-internal-wombat/test-es-index-cleaner-with-prefix updated logger.go:42: 06:58:43 | es-index-cleaner-autoprov/3-install | test step completed 3-install logger.go:42: 06:58:43 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | starting test step 4-wait-es-index-cleaner logger.go:42: 06:58:43 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob test-es-index-cleaner-with-prefix-es-index-cleaner --namespace $NAMESPACE] logger.go:42: 06:58:44 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-22T06:58:44Z" level=debug msg="Checking if the test-es-index-cleaner-with-prefix-es-index-cleaner CronJob exists" logger.go:42: 06:58:44 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-22T06:58:44Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 06:58:44 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-22T06:58:44Z" level=info msg="Cronjob test-es-index-cleaner-with-prefix-es-index-cleaner found successfully" logger.go:42: 06:58:44 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-22T06:58:44Z" level=debug msg="Waiting for the next scheduled job from test-es-index-cleaner-with-prefix-es-index-cleaner cronjob" logger.go:42: 06:58:44 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-22T06:58:44Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 06:58:54 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-22T06:58:54Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 06:59:04 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-22T06:59:04Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 06:59:14 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-01-22T06:59:14Z" level=info msg="Job of owner test-es-index-cleaner-with-prefix-es-index-cleaner succeeded after test-es-index-cleaner-with-prefix-es-index-cleaner 30.005973972s" logger.go:42: 06:59:14 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | test step completed 4-wait-es-index-cleaner logger.go:42: 06:59:14 | es-index-cleaner-autoprov/5-install | starting test step 5-install logger.go:42: 06:59:14 | es-index-cleaner-autoprov/5-install | Jaeger:kuttl-test-internal-wombat/test-es-index-cleaner-with-prefix updated logger.go:42: 06:59:14 | es-index-cleaner-autoprov/5-install | test step completed 5-install logger.go:42: 06:59:14 | es-index-cleaner-autoprov/6-check-indices | starting test step 6-check-indices logger.go:42: 06:59:14 | es-index-cleaner-autoprov/6-check-indices | Job:kuttl-test-internal-wombat/00-check-indices created logger.go:42: 06:59:18 | es-index-cleaner-autoprov/6-check-indices | test step completed 6-check-indices logger.go:42: 06:59:18 | es-index-cleaner-autoprov | es-index-cleaner-autoprov events from ns kuttl-test-internal-wombat: logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:25 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestinternalwombattestesindexcl-1-6bb57c587c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestinternalwombattestesindexcl-1-6fjx92 replicaset-controller logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinternalwombattestesindexcl-1-6fjx92 Binding Scheduled Successfully assigned kuttl-test-internal-wombat/elasticsearch-cdm-kuttltestinternalwombattestesindexcl-1-6fjx92 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:25 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestinternalwombattestesindexcl-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestinternalwombattestesindexcl-1-6bb57c587c to 1 deployment-controller logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinternalwombattestesindexcl-1-6fjx92 AddedInterface Add eth0 [10.131.0.26/23] from ovn-kubernetes logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinternalwombattestesindexcl-1-6fjx92.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:3ad088cbfcfb006d87762e96a70ff989e75d49872954347f7efe34d9529c1b7a" already present on machine kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinternalwombattestesindexcl-1-6fjx92.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinternalwombattestesindexcl-1-6fjx92.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinternalwombattestesindexcl-1-6fjx92.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:911ee3f26ae71eab71bf2a1455d1c2ea871cddc78a51189e3ad2f18f63bd3d91" already present on machine kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinternalwombattestesindexcl-1-6fjx92.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinternalwombattestesindexcl-1-6fjx92.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:36 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestinternalwombattestesindexcl-1-6fjx92.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:41 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestinternalwombattestesindexcl-1-6fjx92.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-5b89968df7-4c7q6 Binding Scheduled Successfully assigned kuttl-test-internal-wombat/test-es-index-cleaner-with-prefix-collector-5b89968df7-4c7q6 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:52 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-collector-5b89968df7 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-collector-5b89968df7-4c7q6 replicaset-controller logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:52 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-collector ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-collector-5b89968df7 to 1 deployment-controller logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-f498ffbd5-jflwh Binding Scheduled Successfully assigned kuttl-test-internal-wombat/test-es-index-cleaner-with-prefix-query-f498ffbd5-jflwh to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:52 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-f498ffbd5 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-f498ffbd5-jflwh replicaset-controller logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:52 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-f498ffbd5 to 1 deployment-controller logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-5b89968df7-4c7q6 AddedInterface Add eth0 [10.128.2.53/23] from ovn-kubernetes logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-5b89968df7-4c7q6.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-5b89968df7-4c7q6.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-5b89968df7-4c7q6.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-f498ffbd5-jflwh AddedInterface Add eth0 [10.129.2.26/23] from ovn-kubernetes logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-f498ffbd5-jflwh.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-f498ffbd5-jflwh.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-f498ffbd5-jflwh.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-f498ffbd5-jflwh.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-f498ffbd5-jflwh.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-f498ffbd5-jflwh.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-f498ffbd5-jflwh.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-f498ffbd5-jflwh.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:57:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-f498ffbd5-jflwh.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7f5fcbb489-qr5vs Binding Scheduled Successfully assigned kuttl-test-internal-wombat/test-es-index-cleaner-with-prefix-query-7f5fcbb489-qr5vs to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:00 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-7f5fcbb489 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-7f5fcbb489-qr5vs replicaset-controller logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-f498ffbd5-jflwh.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-f498ffbd5-jflwh.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-f498ffbd5-jflwh.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:00 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-f498ffbd5 SuccessfulDelete Deleted pod: test-es-index-cleaner-with-prefix-query-f498ffbd5-jflwh replicaset-controller logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:00 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled down replica set test-es-index-cleaner-with-prefix-query-f498ffbd5 to 0 from 1 deployment-controller logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:00 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-7f5fcbb489 to 1 deployment-controller logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7f5fcbb489-qr5vs AddedInterface Add eth0 [10.129.2.27/23] from ovn-kubernetes logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7f5fcbb489-qr5vs.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7f5fcbb489-qr5vs.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7f5fcbb489-qr5vs.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7f5fcbb489-qr5vs.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7f5fcbb489-qr5vs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7f5fcbb489-qr5vs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7f5fcbb489-qr5vs.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7f5fcbb489-qr5vs.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7f5fcbb489-qr5vs.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:04 +0000 UTC Normal Pod 00-report-span-7znt6 Binding Scheduled Successfully assigned kuttl-test-internal-wombat/00-report-span-7znt6 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:04 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-7znt6 job-controller logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:05 +0000 UTC Normal Pod 00-report-span-7znt6 AddedInterface Add eth0 [10.128.2.54/23] from ovn-kubernetes logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:05 +0000 UTC Normal Pod 00-report-span-7znt6.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:05 +0000 UTC Normal Pod 00-report-span-7znt6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:05 +0000 UTC Normal Pod 00-report-span-7znt6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:07 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:08 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:08 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:38 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:38 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-5b89968df7-4c7q6 horizontal-pod-autoscaler logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:38 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:58:42 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:59:00 +0000 UTC Normal Job.batch test-es-index-cleaner-with-prefix-es-index-cleaner-28431779 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-es-index-cleaner-2843177mq7cj job-controller logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:59:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2843177mq7cj Binding Scheduled Successfully assigned kuttl-test-internal-wombat/test-es-index-cleaner-with-prefix-es-index-cleaner-2843177mq7cj to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:59:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2843177mq7cj AddedInterface Add eth0 [10.128.2.55/23] from ovn-kubernetes logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:59:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2843177mq7cj.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:e5bea9a64ae418869cfb556d70e0a586c21589aed8606b4ff5850780ff5bbbd6" kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:59:00 +0000 UTC Normal CronJob.batch test-es-index-cleaner-with-prefix-es-index-cleaner SuccessfulCreate Created job test-es-index-cleaner-with-prefix-es-index-cleaner-28431779 cronjob-controller logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:59:05 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2843177mq7cj.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:e5bea9a64ae418869cfb556d70e0a586c21589aed8606b4ff5850780ff5bbbd6" in 4.915s (4.915s including waiting) kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:59:05 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2843177mq7cj.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Created Created container test-es-index-cleaner-with-prefix-es-index-cleaner kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:59:05 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2843177mq7cj.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Started Started container test-es-index-cleaner-with-prefix-es-index-cleaner kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:59:08 +0000 UTC Normal Job.batch test-es-index-cleaner-with-prefix-es-index-cleaner-28431779 Completed Job completed job-controller logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:59:08 +0000 UTC Normal CronJob.batch test-es-index-cleaner-with-prefix-es-index-cleaner SawCompletedJob Saw completed job: test-es-index-cleaner-with-prefix-es-index-cleaner-28431779, status: Complete cronjob-controller logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:59:14 +0000 UTC Normal Pod 00-check-indices-9xjck Binding Scheduled Successfully assigned kuttl-test-internal-wombat/00-check-indices-9xjck to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:59:14 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-9xjck job-controller logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:59:15 +0000 UTC Normal Pod 00-check-indices-9xjck AddedInterface Add eth0 [10.128.2.56/23] from ovn-kubernetes logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:59:15 +0000 UTC Normal Pod 00-check-indices-9xjck.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:59:15 +0000 UTC Normal Pod 00-check-indices-9xjck.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:59:15 +0000 UTC Normal Pod 00-check-indices-9xjck.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 06:59:18 | es-index-cleaner-autoprov | 2024-01-22 06:59:18 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 06:59:18 | es-index-cleaner-autoprov | Deleting namespace: kuttl-test-internal-wombat === CONT kuttl/harness/es-from-aio-to-production logger.go:42: 06:59:25 | es-from-aio-to-production | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 06:59:25 | es-from-aio-to-production | Creating namespace: kuttl-test-fit-prawn logger.go:42: 06:59:25 | es-from-aio-to-production/0-install | starting test step 0-install logger.go:42: 06:59:25 | es-from-aio-to-production/0-install | Jaeger:kuttl-test-fit-prawn/my-jaeger created logger.go:42: 06:59:36 | es-from-aio-to-production/0-install | test step completed 0-install logger.go:42: 06:59:36 | es-from-aio-to-production/1-smoke-test | starting test step 1-smoke-test logger.go:42: 06:59:36 | es-from-aio-to-production/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 06:59:37 | es-from-aio-to-production/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 06:59:44 | es-from-aio-to-production/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 06:59:44 | es-from-aio-to-production/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 06:59:44 | es-from-aio-to-production/1-smoke-test | job.batch/report-span created logger.go:42: 06:59:44 | es-from-aio-to-production/1-smoke-test | job.batch/check-span created logger.go:42: 06:59:50 | es-from-aio-to-production/1-smoke-test | test step completed 1-smoke-test logger.go:42: 06:59:50 | es-from-aio-to-production/3-install | starting test step 3-install logger.go:42: 06:59:51 | es-from-aio-to-production/3-install | Jaeger:kuttl-test-fit-prawn/my-jaeger updated logger.go:42: 07:00:23 | es-from-aio-to-production/3-install | test step completed 3-install logger.go:42: 07:00:23 | es-from-aio-to-production/4-smoke-test | starting test step 4-smoke-test logger.go:42: 07:00:23 | es-from-aio-to-production/4-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:00:30 | es-from-aio-to-production/4-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:00:31 | es-from-aio-to-production/4-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:00:31 | es-from-aio-to-production/4-smoke-test | job.batch/report-span unchanged logger.go:42: 07:00:31 | es-from-aio-to-production/4-smoke-test | job.batch/check-span unchanged logger.go:42: 07:00:31 | es-from-aio-to-production/4-smoke-test | test step completed 4-smoke-test logger.go:42: 07:00:31 | es-from-aio-to-production | es-from-aio-to-production events from ns kuttl-test-fit-prawn: logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:29 +0000 UTC Normal Pod my-jaeger-88d49fc58-7vplc Binding Scheduled Successfully assigned kuttl-test-fit-prawn/my-jaeger-88d49fc58-7vplc to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:29 +0000 UTC Normal ReplicaSet.apps my-jaeger-88d49fc58 SuccessfulCreate Created pod: my-jaeger-88d49fc58-7vplc replicaset-controller logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:29 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-88d49fc58 to 1 deployment-controller logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:30 +0000 UTC Warning Pod my-jaeger-88d49fc58-7vplc FailedMount MountVolume.SetUp failed for volume "my-jaeger-sampling-configuration-volume" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:31 +0000 UTC Normal Pod my-jaeger-88d49fc58-7vplc AddedInterface Add eth0 [10.131.0.27/23] from ovn-kubernetes logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:31 +0000 UTC Normal Pod my-jaeger-88d49fc58-7vplc.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:34 +0000 UTC Normal Pod my-jaeger-88d49fc58-7vplc.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" in 3.237s (3.237s including waiting) kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:34 +0000 UTC Normal Pod my-jaeger-88d49fc58-7vplc.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:34 +0000 UTC Normal Pod my-jaeger-88d49fc58-7vplc.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:34 +0000 UTC Normal Pod my-jaeger-88d49fc58-7vplc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:34 +0000 UTC Normal Pod my-jaeger-88d49fc58-7vplc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:34 +0000 UTC Normal Pod my-jaeger-88d49fc58-7vplc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:40 +0000 UTC Normal Pod my-jaeger-88d49fc58-7vplc.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:40 +0000 UTC Normal Pod my-jaeger-88d49fc58-7vplc.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:40 +0000 UTC Normal ReplicaSet.apps my-jaeger-88d49fc58 SuccessfulDelete Deleted pod: my-jaeger-88d49fc58-7vplc replicaset-controller logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:40 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-88d49fc58 to 0 from 1 deployment-controller logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:41 +0000 UTC Normal Pod my-jaeger-7f787b5f6c-dl6h6 Binding Scheduled Successfully assigned kuttl-test-fit-prawn/my-jaeger-7f787b5f6c-dl6h6 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:41 +0000 UTC Normal Pod my-jaeger-7f787b5f6c-dl6h6 AddedInterface Add eth0 [10.131.0.28/23] from ovn-kubernetes logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:41 +0000 UTC Normal Pod my-jaeger-7f787b5f6c-dl6h6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:41 +0000 UTC Normal Pod my-jaeger-7f787b5f6c-dl6h6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:41 +0000 UTC Normal Pod my-jaeger-7f787b5f6c-dl6h6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:41 +0000 UTC Normal Pod my-jaeger-7f787b5f6c-dl6h6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:41 +0000 UTC Normal Pod my-jaeger-7f787b5f6c-dl6h6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:41 +0000 UTC Normal Pod my-jaeger-7f787b5f6c-dl6h6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:41 +0000 UTC Normal ReplicaSet.apps my-jaeger-7f787b5f6c SuccessfulCreate Created pod: my-jaeger-7f787b5f6c-dl6h6 replicaset-controller logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:41 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-7f787b5f6c to 1 deployment-controller logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:44 +0000 UTC Normal Pod check-span-2qgxn Binding Scheduled Successfully assigned kuttl-test-fit-prawn/check-span-2qgxn to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:44 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-2qgxn job-controller logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:44 +0000 UTC Normal Pod report-span-hkdj2 Binding Scheduled Successfully assigned kuttl-test-fit-prawn/report-span-hkdj2 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:44 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-hkdj2 job-controller logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:45 +0000 UTC Normal Pod check-span-2qgxn AddedInterface Add eth0 [10.129.2.28/23] from ovn-kubernetes logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:45 +0000 UTC Normal Pod check-span-2qgxn.spec.containers{asserts-container} Pulling Pulling image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:45 +0000 UTC Normal Pod report-span-hkdj2 AddedInterface Add eth0 [10.128.2.57/23] from ovn-kubernetes logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:45 +0000 UTC Normal Pod report-span-hkdj2.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:45 +0000 UTC Normal Pod report-span-hkdj2.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:45 +0000 UTC Normal Pod report-span-hkdj2.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:47 +0000 UTC Normal Pod check-span-2qgxn.spec.containers{asserts-container} Pulled Successfully pulled image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" in 1.481s (1.481s including waiting) kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:47 +0000 UTC Normal Pod check-span-2qgxn.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:47 +0000 UTC Normal Pod check-span-2qgxn.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:50 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfitprawnmyjaeger-1-cf9554677-tj22q Binding Scheduled Successfully assigned kuttl-test-fit-prawn/elasticsearch-cdm-kuttltestfitprawnmyjaeger-1-cf9554677-tj22q to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:54 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestfitprawnmyjaeger-1-cf9554677 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestfitprawnmyjaeger-1-cf9554677-tj22q replicaset-controller logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:54 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestfitprawnmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestfitprawnmyjaeger-1-cf9554677 to 1 deployment-controller logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfitprawnmyjaeger-1-cf9554677-tj22q AddedInterface Add eth0 [10.128.2.58/23] from ovn-kubernetes logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfitprawnmyjaeger-1-cf9554677-tj22q.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:3ad088cbfcfb006d87762e96a70ff989e75d49872954347f7efe34d9529c1b7a" already present on machine kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfitprawnmyjaeger-1-cf9554677-tj22q.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfitprawnmyjaeger-1-cf9554677-tj22q.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfitprawnmyjaeger-1-cf9554677-tj22q.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:911ee3f26ae71eab71bf2a1455d1c2ea871cddc78a51189e3ad2f18f63bd3d91" already present on machine kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfitprawnmyjaeger-1-cf9554677-tj22q.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 06:59:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfitprawnmyjaeger-1-cf9554677-tj22q.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:05 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfitprawnmyjaeger-1-cf9554677-tj22q.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:10 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfitprawnmyjaeger-1-cf9554677-tj22q.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:17 +0000 UTC Normal Job.batch report-span Completed Job completed job-controller logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:21 +0000 UTC Normal Pod my-jaeger-collector-86c779848c-khgw8 Binding Scheduled Successfully assigned kuttl-test-fit-prawn/my-jaeger-collector-86c779848c-khgw8 to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:21 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-86c779848c SuccessfulCreate Created pod: my-jaeger-collector-86c779848c-khgw8 replicaset-controller logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:21 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-86c779848c to 1 deployment-controller logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:21 +0000 UTC Normal Pod my-jaeger-query-b7b9f5fff-dcb8h Binding Scheduled Successfully assigned kuttl-test-fit-prawn/my-jaeger-query-b7b9f5fff-dcb8h to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:21 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-b7b9f5fff SuccessfulCreate Created pod: my-jaeger-query-b7b9f5fff-dcb8h replicaset-controller logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:21 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-b7b9f5fff to 1 deployment-controller logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:22 +0000 UTC Normal Pod my-jaeger-collector-86c779848c-khgw8 AddedInterface Add eth0 [10.129.2.29/23] from ovn-kubernetes logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:22 +0000 UTC Normal Pod my-jaeger-collector-86c779848c-khgw8.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:22 +0000 UTC Normal Pod my-jaeger-collector-86c779848c-khgw8.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:22 +0000 UTC Normal Pod my-jaeger-collector-86c779848c-khgw8.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:22 +0000 UTC Normal Pod my-jaeger-query-b7b9f5fff-dcb8h AddedInterface Add eth0 [10.129.2.30/23] from ovn-kubernetes logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:22 +0000 UTC Normal Pod my-jaeger-query-b7b9f5fff-dcb8h.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:22 +0000 UTC Normal Pod my-jaeger-query-b7b9f5fff-dcb8h.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:22 +0000 UTC Normal Pod my-jaeger-query-b7b9f5fff-dcb8h.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:22 +0000 UTC Normal Pod my-jaeger-query-b7b9f5fff-dcb8h.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:22 +0000 UTC Normal Pod my-jaeger-query-b7b9f5fff-dcb8h.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:22 +0000 UTC Normal Pod my-jaeger-query-b7b9f5fff-dcb8h.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:22 +0000 UTC Normal Pod my-jaeger-query-b7b9f5fff-dcb8h.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:22 +0000 UTC Normal Pod my-jaeger-query-b7b9f5fff-dcb8h.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:22 +0000 UTC Normal Pod my-jaeger-query-b7b9f5fff-dcb8h.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:23 +0000 UTC Normal Pod my-jaeger-7f787b5f6c-dl6h6.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | 2024-01-22 07:00:23 +0000 UTC Normal Pod my-jaeger-7f787b5f6c-dl6h6.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:00:31 | es-from-aio-to-production | Deleting namespace: kuttl-test-fit-prawn === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (682.77s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.87s) --- PASS: kuttl/harness/es-multiinstance (114.22s) --- PASS: kuttl/harness/es-simple-prod (5.81s) --- PASS: kuttl/harness/es-rollover-autoprov (238.57s) --- PASS: kuttl/harness/es-increasing-replicas (119.78s) --- PASS: kuttl/harness/es-index-cleaner-autoprov (125.97s) --- PASS: kuttl/harness/es-from-aio-to-production (72.51s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name elasticsearch --report --output /logs/artifacts/elasticsearch.xml ./artifacts/kuttl-report.xml time="2024-01-22T07:00:40Z" level=debug msg="Setting a new name for the test suites" time="2024-01-22T07:00:40Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-01-22T07:00:40Z" level=debug msg="normalizing test case names" time="2024-01-22T07:00:40Z" level=debug msg="elasticsearch/artifacts -> elasticsearch_artifacts" time="2024-01-22T07:00:40Z" level=debug msg="elasticsearch/es-multiinstance -> elasticsearch_es_multiinstance" time="2024-01-22T07:00:40Z" level=debug msg="elasticsearch/es-simple-prod -> elasticsearch_es_simple_prod" time="2024-01-22T07:00:40Z" level=debug msg="elasticsearch/es-rollover-autoprov -> elasticsearch_es_rollover_autoprov" time="2024-01-22T07:00:40Z" level=debug msg="elasticsearch/es-increasing-replicas -> elasticsearch_es_increasing_replicas" time="2024-01-22T07:00:40Z" level=debug msg="elasticsearch/es-index-cleaner-autoprov -> elasticsearch_es_index_cleaner_autoprov" time="2024-01-22T07:00:40Z" level=debug msg="elasticsearch/es-from-aio-to-production -> elasticsearch_es_from_aio_to_production" +-----------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------+--------+ | elasticsearch_artifacts | passed | | elasticsearch_es_multiinstance | passed | | elasticsearch_es_simple_prod | passed | | elasticsearch_es_rollover_autoprov | passed | | elasticsearch_es_increasing_replicas | passed | | elasticsearch_es_index_cleaner_autoprov | passed | | elasticsearch_es_from_aio_to_production | passed | +-----------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh examples false true + '[' 3 -ne 3 ']' + test_suite_name=examples + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/examples.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-examples make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ VERTX_IMG=jaegertracing/vertx-create-span:operator-e2e-tests \ ./tests/e2e/examples/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-18-050837 True False 21m Cluster version is 4.15.0-0.nightly-2024-01-18-050837' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-18-050837 True False 21m Cluster version is 4.15.0-0.nightly-2024-01-18-050837' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/examples/render.sh ++ export SUITE_DIR=./tests/e2e/examples ++ SUITE_DIR=./tests/e2e/examples ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/examples ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test examples-agent-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-agent-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-with-priority-class\e[0m' Rendering files for test examples-agent-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + mkdir -p examples-agent-with-priority-class + cd examples-agent-with-priority-class + example_name=agent-with-priority-class + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + '[' true '!=' true ']' + render_install_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + install_file=./02-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/agent-with-priority-class.yaml -o ./02-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./02-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./02-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./02-install.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + JAEGER_NAME=agent-as-daemonset + local jaeger_strategy ++ get_jaeger_strategy ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./02-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./02-install.yaml ++ strategy=DaemonSet ++ '[' DaemonSet = null ']' ++ echo DaemonSet ++ return 0 + jaeger_strategy=DaemonSet + '[' DaemonSet = DaemonSet ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./02-assert.yaml + render_smoke_test_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + jaeger_name=agent-as-daemonset + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test agent-as-daemonset true 02 + '[' 3 -ne 3 ']' + jaeger=agent-as-daemonset + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-all-in-one-with-options + '[' 1 -ne 1 ']' + test_name=examples-all-in-one-with-options + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-all-in-one-with-options' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-all-in-one-with-options\e[0m' Rendering files for test examples-all-in-one-with-options + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-with-priority-class + '[' examples-agent-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-all-in-one-with-options + cd examples-all-in-one-with-options + example_name=all-in-one-with-options + render_install_example all-in-one-with-options 00 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/all-in-one-with-options.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + JAEGER_NAME=my-jaeger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.metadata.name="my-jaeger"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i 'del(.spec.allInOne.image)' ./00-install.yaml + render_smoke_test_example all-in-one-with-options 01 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + jaeger_name=my-jaeger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + sed -i s~my-jaeger-query:443~my-jaeger-query:443/jaeger~gi ./01-smoke-test.yaml + start_test examples-business-application-injected-sidecar + '[' 1 -ne 1 ']' + test_name=examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-business-application-injected-sidecar' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-business-application-injected-sidecar\e[0m' Rendering files for test examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-all-in-one-with-options + '[' examples-all-in-one-with-options '!=' _build ']' + cd .. + mkdir -p examples-business-application-injected-sidecar + cd examples-business-application-injected-sidecar + example_name=simplest + cp /tmp/jaeger-tests/examples/business-application-injected-sidecar.yaml ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].image=strenv(VERTX_IMG)' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.path="/"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.port=8080' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.initialDelaySeconds=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.failureThreshold=3' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.periodSeconds=10' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.successThreshold=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.timeoutSeconds=1' ./00-install.yaml + render_install_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example simplest 02 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 02 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-collector-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-collector-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-collector-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-collector-with-priority-class\e[0m' Rendering files for test examples-collector-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-business-application-injected-sidecar + '[' examples-business-application-injected-sidecar '!=' _build ']' + cd .. + mkdir -p examples-collector-with-priority-class + cd examples-collector-with-priority-class + example_name=collector-with-priority-class + render_install_example collector-with-priority-class 00 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/collector-with-priority-class.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + JAEGER_NAME=collector-with-high-priority + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example collector-with-priority-class 01 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + jaeger_name=collector-with-high-priority + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test collector-with-high-priority true 01 + '[' 3 -ne 3 ']' + jaeger=collector-with-high-priority + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + export JAEGER_NAME=collector-with-high-priority + JAEGER_NAME=collector-with-high-priority + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-service-types + '[' 1 -ne 1 ']' + test_name=examples-service-types + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-service-types' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-service-types\e[0m' Rendering files for test examples-service-types + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-collector-with-priority-class + '[' examples-collector-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-service-types + cd examples-service-types + example_name=service-types + render_install_example service-types 00 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/service-types.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + JAEGER_NAME=service-types + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example service-types 01 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/service-types.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/service-types.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/service-types.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/service-types.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + jaeger_name=service-types + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test service-types true 01 + '[' 3 -ne 3 ']' + jaeger=service-types + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + export JAEGER_NAME=service-types + JAEGER_NAME=service-types + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod + '[' 1 -ne 1 ']' + test_name=examples-simple-prod + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod\e[0m' Rendering files for test examples-simple-prod + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-service-types + '[' examples-service-types '!=' _build ']' + cd .. + mkdir -p examples-simple-prod + cd examples-simple-prod + example_name=simple-prod + render_install_example simple-prod 01 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod 02 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod-with-volumes + '[' 1 -ne 1 ']' + test_name=examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod-with-volumes' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod-with-volumes\e[0m' Rendering files for test examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod + '[' examples-simple-prod '!=' _build ']' + cd .. + mkdir -p examples-simple-prod-with-volumes + cd examples-simple-prod-with-volumes + example_name=simple-prod-with-volumes + render_install_example simple-prod-with-volumes 01 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod-with-volumes 02 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + /tmp/jaeger-tests/bin/gomplate -f ./03-check-volume.yaml.template -o 03-check-volume.yaml + start_test examples-simplest + '[' 1 -ne 1 ']' + test_name=examples-simplest + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simplest' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simplest\e[0m' Rendering files for test examples-simplest + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod-with-volumes + '[' examples-simple-prod-with-volumes '!=' _build ']' + cd .. + mkdir -p examples-simplest + cd examples-simplest + example_name=simplest + render_install_example simplest 00 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 01 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger + '[' 1 -ne 1 ']' + test_name=examples-with-badger + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger\e[0m' Rendering files for test examples-with-badger + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simplest + '[' examples-simplest '!=' _build ']' + cd .. + mkdir -p examples-with-badger + cd examples-with-badger + example_name=with-badger + render_install_example with-badger 00 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + JAEGER_NAME=with-badger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger 01 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + jaeger_name=with-badger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + export JAEGER_NAME=with-badger + JAEGER_NAME=with-badger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger-and-volume + '[' 1 -ne 1 ']' + test_name=examples-with-badger-and-volume + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger-and-volume' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger-and-volume\e[0m' Rendering files for test examples-with-badger-and-volume + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger + '[' examples-with-badger '!=' _build ']' + cd .. + mkdir -p examples-with-badger-and-volume + cd examples-with-badger-and-volume + example_name=with-badger-and-volume + render_install_example with-badger-and-volume 00 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger-and-volume.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + JAEGER_NAME=with-badger-and-volume + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger-and-volume 01 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + jaeger_name=with-badger-and-volume + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger-and-volume true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger-and-volume + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + export JAEGER_NAME=with-badger-and-volume + JAEGER_NAME=with-badger-and-volume + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-cassandra + '[' 1 -ne 1 ']' + test_name=examples-with-cassandra + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-cassandra' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-cassandra\e[0m' Rendering files for test examples-with-cassandra + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger-and-volume + '[' examples-with-badger-and-volume '!=' _build ']' + cd .. + mkdir -p examples-with-cassandra + cd examples-with-cassandra + example_name=with-cassandra + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-cassandra 01 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-cassandra.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + JAEGER_NAME=with-cassandra + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-cassandra 02 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-cassandra.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-cassandra.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + jaeger_name=with-cassandra + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-cassandra true 02 + '[' 3 -ne 3 ']' + jaeger=with-cassandra + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + export JAEGER_NAME=with-cassandra + JAEGER_NAME=with-cassandra + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-sampling + '[' 1 -ne 1 ']' + test_name=examples-with-sampling + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-sampling' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-sampling\e[0m' Rendering files for test examples-with-sampling + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-cassandra + '[' examples-with-cassandra '!=' _build ']' + cd .. + mkdir -p examples-with-sampling + cd examples-with-sampling + export example_name=with-sampling + example_name=with-sampling + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-sampling 01 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-sampling.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + JAEGER_NAME=with-sampling + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-sampling 02 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-sampling.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-sampling.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + jaeger_name=with-sampling + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-sampling true 02 + '[' 3 -ne 3 ']' + jaeger=with-sampling + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + export JAEGER_NAME=with-sampling + JAEGER_NAME=with-sampling + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-agent-as-daemonset + '[' 1 -ne 1 ']' + test_name=examples-agent-as-daemonset + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-as-daemonset' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-as-daemonset\e[0m' Rendering files for test examples-agent-as-daemonset + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-sampling + '[' examples-with-sampling '!=' _build ']' + cd .. + mkdir -p examples-agent-as-daemonset + cd examples-agent-as-daemonset + '[' true = true ']' + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/agent-as-daemonset.yaml -o 02-install.yaml + '[' true = true ']' + start_test examples-openshift-with-htpasswd + '[' 1 -ne 1 ']' + test_name=examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-openshift-with-htpasswd' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-openshift-with-htpasswd\e[0m' Rendering files for test examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-as-daemonset + '[' examples-agent-as-daemonset '!=' _build ']' + cd .. + mkdir -p examples-openshift-with-htpasswd + cd examples-openshift-with-htpasswd + export JAEGER_NAME=with-htpasswd + JAEGER_NAME=with-htpasswd + export JAEGER_USERNAME=awesomeuser + JAEGER_USERNAME=awesomeuser + export JAEGER_PASSWORD=awesomepassword + JAEGER_PASSWORD=awesomepassword + export 'JAEGER_USER_PASSWORD_HASH=awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' + JAEGER_USER_PASSWORD_HASH='awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ echo 'awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ base64 + SECRET=YXdlc29tZXVzZXI6e1NIQX11VWRxUFZVeXFOQm1FUlUwUXhqM0tGYVpuanc9Cg== + /tmp/jaeger-tests/bin/gomplate -f ./00-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/with-htpasswd.yaml -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + export 'GET_URL_COMMAND=kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + export 'URL=https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + INSECURE=true + JAEGER_USERNAME= + JAEGER_PASSWORD= + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-unsecured.yaml + JAEGER_USERNAME=wronguser + JAEGER_PASSWORD=wrongpassword + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-check-unauthorized.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./04-check-authorized.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running examples E2E tests' Running examples E2E tests + cd tests/e2e/examples/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2576283944 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 15 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/examples-agent-as-daemonset === PAUSE kuttl/harness/examples-agent-as-daemonset === RUN kuttl/harness/examples-agent-with-priority-class === PAUSE kuttl/harness/examples-agent-with-priority-class === RUN kuttl/harness/examples-all-in-one-with-options === PAUSE kuttl/harness/examples-all-in-one-with-options === RUN kuttl/harness/examples-business-application-injected-sidecar === PAUSE kuttl/harness/examples-business-application-injected-sidecar === RUN kuttl/harness/examples-collector-with-priority-class === PAUSE kuttl/harness/examples-collector-with-priority-class === RUN kuttl/harness/examples-openshift-with-htpasswd === PAUSE kuttl/harness/examples-openshift-with-htpasswd === RUN kuttl/harness/examples-service-types === PAUSE kuttl/harness/examples-service-types === RUN kuttl/harness/examples-simple-prod === PAUSE kuttl/harness/examples-simple-prod === RUN kuttl/harness/examples-simple-prod-with-volumes === PAUSE kuttl/harness/examples-simple-prod-with-volumes === RUN kuttl/harness/examples-simplest === PAUSE kuttl/harness/examples-simplest === RUN kuttl/harness/examples-with-badger === PAUSE kuttl/harness/examples-with-badger === RUN kuttl/harness/examples-with-badger-and-volume === PAUSE kuttl/harness/examples-with-badger-and-volume === RUN kuttl/harness/examples-with-cassandra === PAUSE kuttl/harness/examples-with-cassandra === RUN kuttl/harness/examples-with-sampling === PAUSE kuttl/harness/examples-with-sampling === CONT kuttl/harness/artifacts logger.go:42: 07:01:16 | artifacts | Creating namespace: kuttl-test-wanted-termite logger.go:42: 07:01:16 | artifacts | artifacts events from ns kuttl-test-wanted-termite: logger.go:42: 07:01:16 | artifacts | Deleting namespace: kuttl-test-wanted-termite === CONT kuttl/harness/examples-simple-prod logger.go:42: 07:01:22 | examples-simple-prod | Creating namespace: kuttl-test-mighty-gibbon logger.go:42: 07:01:22 | examples-simple-prod/1-install | starting test step 1-install logger.go:42: 07:01:22 | examples-simple-prod/1-install | Jaeger:kuttl-test-mighty-gibbon/simple-prod created logger.go:42: 07:01:57 | examples-simple-prod/1-install | test step completed 1-install logger.go:42: 07:01:57 | examples-simple-prod/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:01:57 | examples-simple-prod/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:01:59 | examples-simple-prod/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:02:06 | examples-simple-prod/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:02:07 | examples-simple-prod/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:02:07 | examples-simple-prod/2-smoke-test | job.batch/report-span created logger.go:42: 07:02:07 | examples-simple-prod/2-smoke-test | job.batch/check-span created logger.go:42: 07:02:18 | examples-simple-prod/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:02:18 | examples-simple-prod | examples-simple-prod events from ns kuttl-test-mighty-gibbon: logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:28 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestmightygibbonsimpleprod-1-988f96677 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestmightygibbonsimpleprod-1-988f968qf5c replicaset-controller logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmightygibbonsimpleprod-1-988f968qf5c Binding Scheduled Successfully assigned kuttl-test-mighty-gibbon/elasticsearch-cdm-kuttltestmightygibbonsimpleprod-1-988f968qf5c to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmightygibbonsimpleprod-1-988f968qf5c AddedInterface Add eth0 [10.131.0.29/23] from ovn-kubernetes logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmightygibbonsimpleprod-1-988f968qf5c.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:3ad088cbfcfb006d87762e96a70ff989e75d49872954347f7efe34d9529c1b7a" already present on machine kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmightygibbonsimpleprod-1-988f968qf5c.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmightygibbonsimpleprod-1-988f968qf5c.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmightygibbonsimpleprod-1-988f968qf5c.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:911ee3f26ae71eab71bf2a1455d1c2ea871cddc78a51189e3ad2f18f63bd3d91" already present on machine kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmightygibbonsimpleprod-1-988f968qf5c.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:28 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestmightygibbonsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestmightygibbonsimpleprod-1-988f96677 to 1 deployment-controller logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmightygibbonsimpleprod-1-988f968qf5c.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:38 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmightygibbonsimpleprod-1-988f968qf5c.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:43 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmightygibbonsimpleprod-1-988f968qf5c.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:55 +0000 UTC Normal Pod simple-prod-collector-755bfc5df7-nhxvk Binding Scheduled Successfully assigned kuttl-test-mighty-gibbon/simple-prod-collector-755bfc5df7-nhxvk to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:55 +0000 UTC Normal Pod simple-prod-collector-755bfc5df7-nhxvk AddedInterface Add eth0 [10.128.2.60/23] from ovn-kubernetes logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:55 +0000 UTC Normal Pod simple-prod-collector-755bfc5df7-nhxvk.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:55 +0000 UTC Normal Pod simple-prod-collector-755bfc5df7-nhxvk.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:55 +0000 UTC Normal Pod simple-prod-collector-755bfc5df7-nhxvk.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:55 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-755bfc5df7 SuccessfulCreate Created pod: simple-prod-collector-755bfc5df7-nhxvk replicaset-controller logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:55 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-755bfc5df7 to 1 deployment-controller logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:55 +0000 UTC Normal Pod simple-prod-query-74666b6f98-pxhnj Binding Scheduled Successfully assigned kuttl-test-mighty-gibbon/simple-prod-query-74666b6f98-pxhnj to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:55 +0000 UTC Normal Pod simple-prod-query-74666b6f98-pxhnj AddedInterface Add eth0 [10.129.2.31/23] from ovn-kubernetes logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:55 +0000 UTC Normal Pod simple-prod-query-74666b6f98-pxhnj.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:55 +0000 UTC Normal Pod simple-prod-query-74666b6f98-pxhnj.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:55 +0000 UTC Normal Pod simple-prod-query-74666b6f98-pxhnj.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:55 +0000 UTC Normal Pod simple-prod-query-74666b6f98-pxhnj.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:55 +0000 UTC Normal Pod simple-prod-query-74666b6f98-pxhnj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:55 +0000 UTC Normal ReplicaSet.apps simple-prod-query-74666b6f98 SuccessfulCreate Created pod: simple-prod-query-74666b6f98-pxhnj replicaset-controller logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:55 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-74666b6f98 to 1 deployment-controller logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:56 +0000 UTC Normal Pod simple-prod-query-74666b6f98-pxhnj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:56 +0000 UTC Normal Pod simple-prod-query-74666b6f98-pxhnj.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:56 +0000 UTC Normal Pod simple-prod-query-74666b6f98-pxhnj.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:01:56 +0000 UTC Normal Pod simple-prod-query-74666b6f98-pxhnj.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:02 +0000 UTC Normal Pod simple-prod-query-74666b6f98-pxhnj.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:02 +0000 UTC Normal Pod simple-prod-query-74666b6f98-pxhnj.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:02 +0000 UTC Normal Pod simple-prod-query-74666b6f98-pxhnj.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:02 +0000 UTC Normal ReplicaSet.apps simple-prod-query-74666b6f98 SuccessfulDelete Deleted pod: simple-prod-query-74666b6f98-pxhnj replicaset-controller logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:02 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-74666b6f98 to 0 from 1 deployment-controller logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:03 +0000 UTC Normal Pod simple-prod-query-5fb6dc5c5b-fb48d Binding Scheduled Successfully assigned kuttl-test-mighty-gibbon/simple-prod-query-5fb6dc5c5b-fb48d to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:03 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5fb6dc5c5b SuccessfulCreate Created pod: simple-prod-query-5fb6dc5c5b-fb48d replicaset-controller logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:03 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-5fb6dc5c5b to 1 deployment-controller logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:04 +0000 UTC Normal Pod simple-prod-query-5fb6dc5c5b-fb48d AddedInterface Add eth0 [10.129.2.32/23] from ovn-kubernetes logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:04 +0000 UTC Normal Pod simple-prod-query-5fb6dc5c5b-fb48d.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:04 +0000 UTC Normal Pod simple-prod-query-5fb6dc5c5b-fb48d.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:04 +0000 UTC Normal Pod simple-prod-query-5fb6dc5c5b-fb48d.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:04 +0000 UTC Normal Pod simple-prod-query-5fb6dc5c5b-fb48d.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:04 +0000 UTC Normal Pod simple-prod-query-5fb6dc5c5b-fb48d.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:04 +0000 UTC Normal Pod simple-prod-query-5fb6dc5c5b-fb48d.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:04 +0000 UTC Normal Pod simple-prod-query-5fb6dc5c5b-fb48d.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:04 +0000 UTC Normal Pod simple-prod-query-5fb6dc5c5b-fb48d.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:04 +0000 UTC Normal Pod simple-prod-query-5fb6dc5c5b-fb48d.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:07 +0000 UTC Normal Pod check-span-w94t9 Binding Scheduled Successfully assigned kuttl-test-mighty-gibbon/check-span-w94t9 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:07 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-w94t9 job-controller logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:07 +0000 UTC Normal Pod report-span-s8x66 Binding Scheduled Successfully assigned kuttl-test-mighty-gibbon/report-span-s8x66 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:07 +0000 UTC Normal Pod report-span-s8x66 AddedInterface Add eth0 [10.128.2.61/23] from ovn-kubernetes logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:07 +0000 UTC Normal Pod report-span-s8x66.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:07 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-s8x66 job-controller logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:08 +0000 UTC Normal Pod check-span-w94t9 AddedInterface Add eth0 [10.128.2.62/23] from ovn-kubernetes logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:08 +0000 UTC Normal Pod check-span-w94t9.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:08 +0000 UTC Normal Pod check-span-w94t9.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:08 +0000 UTC Normal Pod check-span-w94t9.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:08 +0000 UTC Normal Pod report-span-s8x66.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:08 +0000 UTC Normal Pod report-span-s8x66.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:02:18 | examples-simple-prod | 2024-01-22 07:02:18 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:02:18 | examples-simple-prod | Deleting namespace: kuttl-test-mighty-gibbon === CONT kuttl/harness/examples-with-sampling logger.go:42: 07:02:30 | examples-with-sampling | Creating namespace: kuttl-test-nice-caribou logger.go:42: 07:02:30 | examples-with-sampling/0-install | starting test step 0-install logger.go:42: 07:02:30 | examples-with-sampling/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:02:30 | examples-with-sampling/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:02:30 | examples-with-sampling/0-install | >>>> Creating namespace kuttl-test-nice-caribou logger.go:42: 07:02:30 | examples-with-sampling/0-install | kubectl create namespace kuttl-test-nice-caribou 2>&1 | grep -v "already exists" || true logger.go:42: 07:02:30 | examples-with-sampling/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-nice-caribou 2>&1 | grep -v "already exists" || true logger.go:42: 07:02:31 | examples-with-sampling/0-install | service/cassandra created logger.go:42: 07:02:31 | examples-with-sampling/0-install | statefulset.apps/cassandra created logger.go:42: 07:02:31 | examples-with-sampling/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:02:43 | examples-with-sampling/0-install | test step completed 0-install logger.go:42: 07:02:43 | examples-with-sampling/1-install | starting test step 1-install logger.go:42: 07:02:43 | examples-with-sampling/1-install | Jaeger:kuttl-test-nice-caribou/with-sampling created logger.go:42: 07:02:49 | examples-with-sampling/1-install | test step completed 1-install logger.go:42: 07:02:49 | examples-with-sampling/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:02:49 | examples-with-sampling/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-sampling /dev/null] logger.go:42: 07:02:50 | examples-with-sampling/2-smoke-test | Warning: resource jaegers/with-sampling is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:02:56 | examples-with-sampling/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:02:57 | examples-with-sampling/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:02:57 | examples-with-sampling/2-smoke-test | job.batch/report-span created logger.go:42: 07:02:57 | examples-with-sampling/2-smoke-test | job.batch/check-span created logger.go:42: 07:03:08 | examples-with-sampling/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:03:08 | examples-with-sampling/3- | starting test step 3- logger.go:42: 07:03:08 | examples-with-sampling/3- | test step completed 3- logger.go:42: 07:03:08 | examples-with-sampling | examples-with-sampling events from ns kuttl-test-nice-caribou: logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:31 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-nice-caribou/cassandra-0 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:31 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.131.0.30/23] from ovn-kubernetes logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:31 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:31 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:36 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 4.165s (4.165s including waiting) kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:36 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:36 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:36 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-nice-caribou/cassandra-1 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:36 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:37 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.128.2.63/23] from ovn-kubernetes logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:37 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:41 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 4.318s (4.318s including waiting) kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:41 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:41 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:46 +0000 UTC Normal Pod with-sampling-5f8647bb8d-nspwj Binding Scheduled Successfully assigned kuttl-test-nice-caribou/with-sampling-5f8647bb8d-nspwj to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:46 +0000 UTC Normal ReplicaSet.apps with-sampling-5f8647bb8d SuccessfulCreate Created pod: with-sampling-5f8647bb8d-nspwj replicaset-controller logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:46 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-5f8647bb8d to 1 deployment-controller logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:47 +0000 UTC Normal Pod with-sampling-5f8647bb8d-nspwj AddedInterface Add eth0 [10.131.0.31/23] from ovn-kubernetes logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:47 +0000 UTC Normal Pod with-sampling-5f8647bb8d-nspwj.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:47 +0000 UTC Normal Pod with-sampling-5f8647bb8d-nspwj.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:47 +0000 UTC Normal Pod with-sampling-5f8647bb8d-nspwj.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:47 +0000 UTC Normal Pod with-sampling-5f8647bb8d-nspwj.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:47 +0000 UTC Normal Pod with-sampling-5f8647bb8d-nspwj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:47 +0000 UTC Normal Pod with-sampling-5f8647bb8d-nspwj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:53 +0000 UTC Normal Pod with-sampling-5f8647bb8d-nspwj.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:53 +0000 UTC Normal Pod with-sampling-5f8647bb8d-nspwj.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:53 +0000 UTC Normal ReplicaSet.apps with-sampling-5f8647bb8d SuccessfulDelete Deleted pod: with-sampling-5f8647bb8d-nspwj replicaset-controller logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:53 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled down replica set with-sampling-5f8647bb8d to 0 from 1 deployment-controller logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:54 +0000 UTC Normal Pod with-sampling-5dfc744478-cfw52 Binding Scheduled Successfully assigned kuttl-test-nice-caribou/with-sampling-5dfc744478-cfw52 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:54 +0000 UTC Normal ReplicaSet.apps with-sampling-5dfc744478 SuccessfulCreate Created pod: with-sampling-5dfc744478-cfw52 replicaset-controller logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:54 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-5dfc744478 to 1 deployment-controller logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:55 +0000 UTC Normal Pod with-sampling-5dfc744478-cfw52 AddedInterface Add eth0 [10.131.0.32/23] from ovn-kubernetes logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:55 +0000 UTC Normal Pod with-sampling-5dfc744478-cfw52.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:55 +0000 UTC Normal Pod with-sampling-5dfc744478-cfw52.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:55 +0000 UTC Normal Pod with-sampling-5dfc744478-cfw52.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:55 +0000 UTC Normal Pod with-sampling-5dfc744478-cfw52.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:55 +0000 UTC Normal Pod with-sampling-5dfc744478-cfw52.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:55 +0000 UTC Normal Pod with-sampling-5dfc744478-cfw52.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:57 +0000 UTC Normal Pod check-span-74wmp Binding Scheduled Successfully assigned kuttl-test-nice-caribou/check-span-74wmp to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:57 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-74wmp job-controller logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:57 +0000 UTC Normal Pod report-span-cdf5k Binding Scheduled Successfully assigned kuttl-test-nice-caribou/report-span-cdf5k to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:57 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-cdf5k job-controller logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:58 +0000 UTC Normal Pod check-span-74wmp AddedInterface Add eth0 [10.128.2.64/23] from ovn-kubernetes logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:58 +0000 UTC Normal Pod check-span-74wmp.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:58 +0000 UTC Normal Pod check-span-74wmp.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:58 +0000 UTC Normal Pod check-span-74wmp.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:58 +0000 UTC Normal Pod report-span-cdf5k AddedInterface Add eth0 [10.129.2.33/23] from ovn-kubernetes logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:58 +0000 UTC Normal Pod report-span-cdf5k.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:58 +0000 UTC Normal Pod report-span-cdf5k.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:02:58 +0000 UTC Normal Pod report-span-cdf5k.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:03:08 | examples-with-sampling | 2024-01-22 07:03:08 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:03:08 | examples-with-sampling | Deleting namespace: kuttl-test-nice-caribou === CONT kuttl/harness/examples-with-cassandra logger.go:42: 07:07:43 | examples-with-cassandra | Creating namespace: kuttl-test-vital-chamois logger.go:42: 07:07:43 | examples-with-cassandra/0-install | starting test step 0-install logger.go:42: 07:07:43 | examples-with-cassandra/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:07:43 | examples-with-cassandra/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:07:43 | examples-with-cassandra/0-install | >>>> Creating namespace kuttl-test-vital-chamois logger.go:42: 07:07:43 | examples-with-cassandra/0-install | kubectl create namespace kuttl-test-vital-chamois 2>&1 | grep -v "already exists" || true logger.go:42: 07:07:43 | examples-with-cassandra/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-vital-chamois 2>&1 | grep -v "already exists" || true logger.go:42: 07:07:43 | examples-with-cassandra/0-install | service/cassandra created logger.go:42: 07:07:43 | examples-with-cassandra/0-install | statefulset.apps/cassandra created logger.go:42: 07:07:43 | examples-with-cassandra/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:07:47 | examples-with-cassandra/0-install | test step completed 0-install logger.go:42: 07:07:47 | examples-with-cassandra/1-install | starting test step 1-install logger.go:42: 07:07:47 | examples-with-cassandra/1-install | Jaeger:kuttl-test-vital-chamois/with-cassandra created logger.go:42: 07:08:08 | examples-with-cassandra/1-install | test step completed 1-install logger.go:42: 07:08:08 | examples-with-cassandra/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:08:08 | examples-with-cassandra/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-cassandra /dev/null] logger.go:42: 07:08:09 | examples-with-cassandra/2-smoke-test | Warning: resource jaegers/with-cassandra is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:08:15 | examples-with-cassandra/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:08:16 | examples-with-cassandra/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:08:16 | examples-with-cassandra/2-smoke-test | job.batch/report-span created logger.go:42: 07:08:16 | examples-with-cassandra/2-smoke-test | job.batch/check-span created logger.go:42: 07:08:28 | examples-with-cassandra/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:08:28 | examples-with-cassandra | examples-with-cassandra events from ns kuttl-test-vital-chamois: logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:44 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-vital-chamois/cassandra-0 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:44 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.131.0.33/23] from ovn-kubernetes logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:44 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:44 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:45 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:45 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:45 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-vital-chamois/cassandra-1 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:45 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.128.2.65/23] from ovn-kubernetes logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:45 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:45 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:46 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:46 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:50 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-xckls Binding Scheduled Successfully assigned kuttl-test-vital-chamois/with-cassandra-cassandra-schema-job-xckls to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:50 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-xckls AddedInterface Add eth0 [10.131.0.34/23] from ovn-kubernetes logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:50 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-xckls.spec.containers{with-cassandra-cassandra-schema-job} Pulling Pulling image "jaegertracing/jaeger-cassandra-schema:1.51.0" kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:50 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job SuccessfulCreate Created pod: with-cassandra-cassandra-schema-job-xckls job-controller logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:55 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-xckls.spec.containers{with-cassandra-cassandra-schema-job} Pulled Successfully pulled image "jaegertracing/jaeger-cassandra-schema:1.51.0" in 4.993s (4.993s including waiting) kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:56 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-xckls.spec.containers{with-cassandra-cassandra-schema-job} Created Created container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:07:56 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-xckls.spec.containers{with-cassandra-cassandra-schema-job} Started Started container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:03 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job Completed Job completed job-controller logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:04 +0000 UTC Normal Pod with-cassandra-68bc7cc99-zl96j Binding Scheduled Successfully assigned kuttl-test-vital-chamois/with-cassandra-68bc7cc99-zl96j to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:04 +0000 UTC Warning Pod with-cassandra-68bc7cc99-zl96j FailedMount MountVolume.SetUp failed for volume "with-cassandra-ui-oauth-proxy-tls" : secret "with-cassandra-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:04 +0000 UTC Normal ReplicaSet.apps with-cassandra-68bc7cc99 SuccessfulCreate Created pod: with-cassandra-68bc7cc99-zl96j replicaset-controller logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:04 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-68bc7cc99 to 1 deployment-controller logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:05 +0000 UTC Normal Pod with-cassandra-68bc7cc99-zl96j AddedInterface Add eth0 [10.131.0.35/23] from ovn-kubernetes logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:05 +0000 UTC Normal Pod with-cassandra-68bc7cc99-zl96j.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:05 +0000 UTC Normal Pod with-cassandra-68bc7cc99-zl96j.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:05 +0000 UTC Normal Pod with-cassandra-68bc7cc99-zl96j.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:05 +0000 UTC Normal Pod with-cassandra-68bc7cc99-zl96j.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:06 +0000 UTC Normal Pod with-cassandra-68bc7cc99-zl96j.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:06 +0000 UTC Normal Pod with-cassandra-68bc7cc99-zl96j.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:10 +0000 UTC Normal Pod with-cassandra-68bc7cc99-zl96j.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:10 +0000 UTC Normal Pod with-cassandra-68bc7cc99-zl96j.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:10 +0000 UTC Normal ReplicaSet.apps with-cassandra-68bc7cc99 SuccessfulDelete Deleted pod: with-cassandra-68bc7cc99-zl96j replicaset-controller logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:10 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled down replica set with-cassandra-68bc7cc99 to 0 from 1 deployment-controller logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:11 +0000 UTC Normal Pod with-cassandra-b5d467d67-mnwwl Binding Scheduled Successfully assigned kuttl-test-vital-chamois/with-cassandra-b5d467d67-mnwwl to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:11 +0000 UTC Normal ReplicaSet.apps with-cassandra-b5d467d67 SuccessfulCreate Created pod: with-cassandra-b5d467d67-mnwwl replicaset-controller logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:11 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-b5d467d67 to 1 deployment-controller logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:12 +0000 UTC Normal Pod with-cassandra-b5d467d67-mnwwl AddedInterface Add eth0 [10.131.0.36/23] from ovn-kubernetes logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:12 +0000 UTC Normal Pod with-cassandra-b5d467d67-mnwwl.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:12 +0000 UTC Normal Pod with-cassandra-b5d467d67-mnwwl.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:12 +0000 UTC Normal Pod with-cassandra-b5d467d67-mnwwl.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:12 +0000 UTC Normal Pod with-cassandra-b5d467d67-mnwwl.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:12 +0000 UTC Normal Pod with-cassandra-b5d467d67-mnwwl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:12 +0000 UTC Normal Pod with-cassandra-b5d467d67-mnwwl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:16 +0000 UTC Normal Pod check-span-5tw7z Binding Scheduled Successfully assigned kuttl-test-vital-chamois/check-span-5tw7z to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:16 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-5tw7z job-controller logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:16 +0000 UTC Normal Pod report-span-4rblx Binding Scheduled Successfully assigned kuttl-test-vital-chamois/report-span-4rblx to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:16 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-4rblx job-controller logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:17 +0000 UTC Normal Pod check-span-5tw7z AddedInterface Add eth0 [10.128.2.66/23] from ovn-kubernetes logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:17 +0000 UTC Normal Pod check-span-5tw7z.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:17 +0000 UTC Normal Pod check-span-5tw7z.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:17 +0000 UTC Normal Pod check-span-5tw7z.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:17 +0000 UTC Normal Pod report-span-4rblx AddedInterface Add eth0 [10.129.2.34/23] from ovn-kubernetes logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:17 +0000 UTC Normal Pod report-span-4rblx.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:17 +0000 UTC Normal Pod report-span-4rblx.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:17 +0000 UTC Normal Pod report-span-4rblx.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:08:28 | examples-with-cassandra | 2024-01-22 07:08:28 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:08:28 | examples-with-cassandra | Deleting namespace: kuttl-test-vital-chamois === CONT kuttl/harness/examples-with-badger-and-volume logger.go:42: 07:08:40 | examples-with-badger-and-volume | Creating namespace: kuttl-test-balanced-impala logger.go:42: 07:08:40 | examples-with-badger-and-volume/0-install | starting test step 0-install logger.go:42: 07:08:40 | examples-with-badger-and-volume/0-install | Jaeger:kuttl-test-balanced-impala/with-badger-and-volume created logger.go:42: 07:08:46 | examples-with-badger-and-volume/0-install | test step completed 0-install logger.go:42: 07:08:46 | examples-with-badger-and-volume/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:08:46 | examples-with-badger-and-volume/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger-and-volume /dev/null] logger.go:42: 07:08:48 | examples-with-badger-and-volume/1-smoke-test | Warning: resource jaegers/with-badger-and-volume is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:08:54 | examples-with-badger-and-volume/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:08:55 | examples-with-badger-and-volume/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:08:55 | examples-with-badger-and-volume/1-smoke-test | job.batch/report-span created logger.go:42: 07:08:55 | examples-with-badger-and-volume/1-smoke-test | job.batch/check-span created logger.go:42: 07:09:08 | examples-with-badger-and-volume/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:09:08 | examples-with-badger-and-volume | examples-with-badger-and-volume events from ns kuttl-test-balanced-impala: logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:44 +0000 UTC Normal Pod with-badger-and-volume-78f96cf7d8-z26vt Binding Scheduled Successfully assigned kuttl-test-balanced-impala/with-badger-and-volume-78f96cf7d8-z26vt to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:44 +0000 UTC Normal ReplicaSet.apps with-badger-and-volume-78f96cf7d8 SuccessfulCreate Created pod: with-badger-and-volume-78f96cf7d8-z26vt replicaset-controller logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:44 +0000 UTC Normal Deployment.apps with-badger-and-volume ScalingReplicaSet Scaled up replica set with-badger-and-volume-78f96cf7d8 to 1 deployment-controller logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:45 +0000 UTC Normal Pod with-badger-and-volume-78f96cf7d8-z26vt AddedInterface Add eth0 [10.131.0.37/23] from ovn-kubernetes logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:45 +0000 UTC Normal Pod with-badger-and-volume-78f96cf7d8-z26vt.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:45 +0000 UTC Normal Pod with-badger-and-volume-78f96cf7d8-z26vt.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:45 +0000 UTC Normal Pod with-badger-and-volume-78f96cf7d8-z26vt.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:45 +0000 UTC Normal Pod with-badger-and-volume-78f96cf7d8-z26vt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:45 +0000 UTC Normal Pod with-badger-and-volume-78f96cf7d8-z26vt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:45 +0000 UTC Normal Pod with-badger-and-volume-78f96cf7d8-z26vt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:51 +0000 UTC Normal Pod with-badger-and-volume-78f96cf7d8-z26vt.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:51 +0000 UTC Normal Pod with-badger-and-volume-78f96cf7d8-z26vt.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:51 +0000 UTC Normal ReplicaSet.apps with-badger-and-volume-78f96cf7d8 SuccessfulDelete Deleted pod: with-badger-and-volume-78f96cf7d8-z26vt replicaset-controller logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:51 +0000 UTC Normal Deployment.apps with-badger-and-volume ScalingReplicaSet Scaled down replica set with-badger-and-volume-78f96cf7d8 to 0 from 1 deployment-controller logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:52 +0000 UTC Normal Pod with-badger-and-volume-945bb588d-wj8vt Binding Scheduled Successfully assigned kuttl-test-balanced-impala/with-badger-and-volume-945bb588d-wj8vt to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:52 +0000 UTC Normal ReplicaSet.apps with-badger-and-volume-945bb588d SuccessfulCreate Created pod: with-badger-and-volume-945bb588d-wj8vt replicaset-controller logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:52 +0000 UTC Normal Deployment.apps with-badger-and-volume ScalingReplicaSet Scaled up replica set with-badger-and-volume-945bb588d to 1 deployment-controller logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:53 +0000 UTC Normal Pod with-badger-and-volume-945bb588d-wj8vt AddedInterface Add eth0 [10.131.0.38/23] from ovn-kubernetes logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:53 +0000 UTC Normal Pod with-badger-and-volume-945bb588d-wj8vt.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:53 +0000 UTC Normal Pod with-badger-and-volume-945bb588d-wj8vt.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:53 +0000 UTC Normal Pod with-badger-and-volume-945bb588d-wj8vt.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:53 +0000 UTC Normal Pod with-badger-and-volume-945bb588d-wj8vt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:53 +0000 UTC Normal Pod with-badger-and-volume-945bb588d-wj8vt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:53 +0000 UTC Normal Pod with-badger-and-volume-945bb588d-wj8vt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:55 +0000 UTC Normal Pod check-span-95m5t Binding Scheduled Successfully assigned kuttl-test-balanced-impala/check-span-95m5t to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:55 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-95m5t job-controller logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:55 +0000 UTC Normal Pod report-span-lwwl6 Binding Scheduled Successfully assigned kuttl-test-balanced-impala/report-span-lwwl6 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:55 +0000 UTC Normal Pod report-span-lwwl6 AddedInterface Add eth0 [10.128.2.67/23] from ovn-kubernetes logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:55 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-lwwl6 job-controller logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:56 +0000 UTC Normal Pod check-span-95m5t AddedInterface Add eth0 [10.129.2.35/23] from ovn-kubernetes logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:56 +0000 UTC Normal Pod check-span-95m5t.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:56 +0000 UTC Normal Pod check-span-95m5t.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:56 +0000 UTC Normal Pod check-span-95m5t.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:56 +0000 UTC Normal Pod report-span-lwwl6.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:56 +0000 UTC Normal Pod report-span-lwwl6.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:08:56 +0000 UTC Normal Pod report-span-lwwl6.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:09:08 | examples-with-badger-and-volume | 2024-01-22 07:09:07 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:09:08 | examples-with-badger-and-volume | Deleting namespace: kuttl-test-balanced-impala === CONT kuttl/harness/examples-with-badger logger.go:42: 07:09:20 | examples-with-badger | Creating namespace: kuttl-test-equal-rabbit logger.go:42: 07:09:20 | examples-with-badger/0-install | starting test step 0-install logger.go:42: 07:09:20 | examples-with-badger/0-install | Jaeger:kuttl-test-equal-rabbit/with-badger created logger.go:42: 07:09:26 | examples-with-badger/0-install | test step completed 0-install logger.go:42: 07:09:26 | examples-with-badger/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:09:26 | examples-with-badger/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger /dev/null] logger.go:42: 07:09:29 | examples-with-badger/1-smoke-test | Warning: resource jaegers/with-badger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:09:36 | examples-with-badger/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:09:37 | examples-with-badger/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:09:37 | examples-with-badger/1-smoke-test | job.batch/report-span created logger.go:42: 07:09:37 | examples-with-badger/1-smoke-test | job.batch/check-span created logger.go:42: 07:09:48 | examples-with-badger/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:09:48 | examples-with-badger | examples-with-badger events from ns kuttl-test-equal-rabbit: logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:24 +0000 UTC Normal Pod with-badger-8584458945-vd78b Binding Scheduled Successfully assigned kuttl-test-equal-rabbit/with-badger-8584458945-vd78b to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:24 +0000 UTC Normal ReplicaSet.apps with-badger-8584458945 SuccessfulCreate Created pod: with-badger-8584458945-vd78b replicaset-controller logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:24 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-8584458945 to 1 deployment-controller logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:25 +0000 UTC Normal Pod with-badger-8584458945-vd78b AddedInterface Add eth0 [10.131.0.39/23] from ovn-kubernetes logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:25 +0000 UTC Normal Pod with-badger-8584458945-vd78b.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:25 +0000 UTC Normal Pod with-badger-8584458945-vd78b.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:25 +0000 UTC Normal Pod with-badger-8584458945-vd78b.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:25 +0000 UTC Normal Pod with-badger-8584458945-vd78b.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:25 +0000 UTC Normal Pod with-badger-8584458945-vd78b.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:25 +0000 UTC Normal Pod with-badger-8584458945-vd78b.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:33 +0000 UTC Normal Pod with-badger-8584458945-vd78b.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:33 +0000 UTC Normal Pod with-badger-8584458945-vd78b.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:33 +0000 UTC Normal ReplicaSet.apps with-badger-8584458945 SuccessfulDelete Deleted pod: with-badger-8584458945-vd78b replicaset-controller logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:33 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled down replica set with-badger-8584458945 to 0 from 1 deployment-controller logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:34 +0000 UTC Normal Pod with-badger-54877577d7-89qds Binding Scheduled Successfully assigned kuttl-test-equal-rabbit/with-badger-54877577d7-89qds to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:34 +0000 UTC Normal ReplicaSet.apps with-badger-54877577d7 SuccessfulCreate Created pod: with-badger-54877577d7-89qds replicaset-controller logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:34 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-54877577d7 to 1 deployment-controller logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:35 +0000 UTC Normal Pod with-badger-54877577d7-89qds AddedInterface Add eth0 [10.131.0.40/23] from ovn-kubernetes logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:35 +0000 UTC Normal Pod with-badger-54877577d7-89qds.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:35 +0000 UTC Normal Pod with-badger-54877577d7-89qds.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:35 +0000 UTC Normal Pod with-badger-54877577d7-89qds.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:35 +0000 UTC Normal Pod with-badger-54877577d7-89qds.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:35 +0000 UTC Normal Pod with-badger-54877577d7-89qds.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:35 +0000 UTC Normal Pod with-badger-54877577d7-89qds.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:37 +0000 UTC Normal Pod check-span-n7bbr Binding Scheduled Successfully assigned kuttl-test-equal-rabbit/check-span-n7bbr to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:37 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-n7bbr job-controller logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:37 +0000 UTC Normal Pod report-span-d5tf6 Binding Scheduled Successfully assigned kuttl-test-equal-rabbit/report-span-d5tf6 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:37 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-d5tf6 job-controller logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:38 +0000 UTC Normal Pod check-span-n7bbr AddedInterface Add eth0 [10.129.2.36/23] from ovn-kubernetes logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:38 +0000 UTC Normal Pod check-span-n7bbr.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:38 +0000 UTC Normal Pod check-span-n7bbr.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:38 +0000 UTC Normal Pod check-span-n7bbr.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:38 +0000 UTC Normal Pod report-span-d5tf6 AddedInterface Add eth0 [10.128.2.68/23] from ovn-kubernetes logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:38 +0000 UTC Normal Pod report-span-d5tf6.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:38 +0000 UTC Normal Pod report-span-d5tf6.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:38 +0000 UTC Normal Pod report-span-d5tf6.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:09:48 | examples-with-badger | 2024-01-22 07:09:48 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:09:48 | examples-with-badger | Deleting namespace: kuttl-test-equal-rabbit === CONT kuttl/harness/examples-simplest logger.go:42: 07:10:01 | examples-simplest | Creating namespace: kuttl-test-national-panda logger.go:42: 07:10:01 | examples-simplest/0-install | starting test step 0-install logger.go:42: 07:10:01 | examples-simplest/0-install | Jaeger:kuttl-test-national-panda/simplest created logger.go:42: 07:10:08 | examples-simplest/0-install | test step completed 0-install logger.go:42: 07:10:08 | examples-simplest/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:10:08 | examples-simplest/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:10:09 | examples-simplest/1-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:10:15 | examples-simplest/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:10:16 | examples-simplest/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:10:16 | examples-simplest/1-smoke-test | job.batch/report-span created logger.go:42: 07:10:16 | examples-simplest/1-smoke-test | job.batch/check-span created logger.go:42: 07:10:28 | examples-simplest/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:10:28 | examples-simplest | examples-simplest events from ns kuttl-test-national-panda: logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:04 +0000 UTC Normal Pod simplest-5c5b4c6f5d-czrhh Binding Scheduled Successfully assigned kuttl-test-national-panda/simplest-5c5b4c6f5d-czrhh to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:04 +0000 UTC Normal ReplicaSet.apps simplest-5c5b4c6f5d SuccessfulCreate Created pod: simplest-5c5b4c6f5d-czrhh replicaset-controller logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:04 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-5c5b4c6f5d to 1 deployment-controller logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:05 +0000 UTC Normal Pod simplest-5c5b4c6f5d-czrhh AddedInterface Add eth0 [10.131.0.41/23] from ovn-kubernetes logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:05 +0000 UTC Normal Pod simplest-5c5b4c6f5d-czrhh.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:05 +0000 UTC Normal Pod simplest-5c5b4c6f5d-czrhh.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:05 +0000 UTC Normal Pod simplest-5c5b4c6f5d-czrhh.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:05 +0000 UTC Normal Pod simplest-5c5b4c6f5d-czrhh.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:06 +0000 UTC Normal Pod simplest-5c5b4c6f5d-czrhh.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:06 +0000 UTC Normal Pod simplest-5c5b4c6f5d-czrhh.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:11 +0000 UTC Normal Pod simplest-5c5b4c6f5d-czrhh.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:11 +0000 UTC Normal ReplicaSet.apps simplest-5c5b4c6f5d SuccessfulDelete Deleted pod: simplest-5c5b4c6f5d-czrhh replicaset-controller logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:11 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-5c5b4c6f5d to 0 from 1 deployment-controller logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:12 +0000 UTC Normal Pod simplest-5c5b4c6f5d-czrhh.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:12 +0000 UTC Normal Pod simplest-65c48d7df4-p9qpt Binding Scheduled Successfully assigned kuttl-test-national-panda/simplest-65c48d7df4-p9qpt to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:12 +0000 UTC Normal ReplicaSet.apps simplest-65c48d7df4 SuccessfulCreate Created pod: simplest-65c48d7df4-p9qpt replicaset-controller logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:12 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-65c48d7df4 to 1 deployment-controller logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:13 +0000 UTC Normal Pod simplest-65c48d7df4-p9qpt AddedInterface Add eth0 [10.131.0.42/23] from ovn-kubernetes logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:13 +0000 UTC Normal Pod simplest-65c48d7df4-p9qpt.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:13 +0000 UTC Normal Pod simplest-65c48d7df4-p9qpt.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:13 +0000 UTC Normal Pod simplest-65c48d7df4-p9qpt.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:13 +0000 UTC Normal Pod simplest-65c48d7df4-p9qpt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:13 +0000 UTC Normal Pod simplest-65c48d7df4-p9qpt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:13 +0000 UTC Normal Pod simplest-65c48d7df4-p9qpt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:16 +0000 UTC Normal Pod check-span-754w2 Binding Scheduled Successfully assigned kuttl-test-national-panda/check-span-754w2 to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:16 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-754w2 job-controller logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:16 +0000 UTC Normal Pod report-span-4t4h8 Binding Scheduled Successfully assigned kuttl-test-national-panda/report-span-4t4h8 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:16 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-4t4h8 job-controller logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:17 +0000 UTC Normal Pod check-span-754w2 AddedInterface Add eth0 [10.129.2.37/23] from ovn-kubernetes logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:17 +0000 UTC Normal Pod check-span-754w2.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:17 +0000 UTC Normal Pod check-span-754w2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:17 +0000 UTC Normal Pod check-span-754w2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:17 +0000 UTC Normal Pod report-span-4t4h8 AddedInterface Add eth0 [10.128.2.69/23] from ovn-kubernetes logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:17 +0000 UTC Normal Pod report-span-4t4h8.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:17 +0000 UTC Normal Pod report-span-4t4h8.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:17 +0000 UTC Normal Pod report-span-4t4h8.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:10:28 | examples-simplest | 2024-01-22 07:10:27 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:10:28 | examples-simplest | Deleting namespace: kuttl-test-national-panda === CONT kuttl/harness/examples-simple-prod-with-volumes logger.go:42: 07:10:35 | examples-simple-prod-with-volumes | Ignoring 03-check-volume.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:10:35 | examples-simple-prod-with-volumes | Creating namespace: kuttl-test-suitable-corgi logger.go:42: 07:10:35 | examples-simple-prod-with-volumes/1-install | starting test step 1-install logger.go:42: 07:10:35 | examples-simple-prod-with-volumes/1-install | Jaeger:kuttl-test-suitable-corgi/simple-prod created logger.go:42: 07:11:12 | examples-simple-prod-with-volumes/1-install | test step completed 1-install logger.go:42: 07:11:12 | examples-simple-prod-with-volumes/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:11:12 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:11:13 | examples-simple-prod-with-volumes/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:11:19 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:11:20 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:11:20 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/report-span created logger.go:42: 07:11:20 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/check-span created logger.go:42: 07:11:32 | examples-simple-prod-with-volumes/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:11:32 | examples-simple-prod-with-volumes/3-check-volume | starting test step 3-check-volume logger.go:42: 07:11:32 | examples-simple-prod-with-volumes/3-check-volume | running command: [sh -c kubectl exec $(kubectl get pods -n $NAMESPACE -l app=jaeger -l app.kubernetes.io/component=collector -o yaml | /tmp/jaeger-tests/bin/yq e '.items[0].metadata.name') -n $NAMESPACE -- ls /usr/share/elasticsearch/data] logger.go:42: 07:11:32 | examples-simple-prod-with-volumes/3-check-volume | test step completed 3-check-volume logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | examples-simple-prod-with-volumes events from ns kuttl-test-suitable-corgi: logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:10:41 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestsuitablecorgisimpleprod-1-fc8ddc7d5 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestsuitablecorgisimpleprod-1-fc8ddh9kr9 replicaset-controller logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:10:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsuitablecorgisimpleprod-1-fc8ddh9kr9 Binding Scheduled Successfully assigned kuttl-test-suitable-corgi/elasticsearch-cdm-kuttltestsuitablecorgisimpleprod-1-fc8ddh9kr9 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:10:41 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsuitablecorgisimpleprod-1-fc8ddh9kr9 FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:10:41 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestsuitablecorgisimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestsuitablecorgisimpleprod-1-fc8ddc7d5 to 1 deployment-controller logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:10:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsuitablecorgisimpleprod-1-fc8ddh9kr9 AddedInterface Add eth0 [10.131.0.43/23] from ovn-kubernetes logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:10:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsuitablecorgisimpleprod-1-fc8ddh9kr9.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:3ad088cbfcfb006d87762e96a70ff989e75d49872954347f7efe34d9529c1b7a" already present on machine kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:10:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsuitablecorgisimpleprod-1-fc8ddh9kr9.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:10:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsuitablecorgisimpleprod-1-fc8ddh9kr9.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:10:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsuitablecorgisimpleprod-1-fc8ddh9kr9.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:911ee3f26ae71eab71bf2a1455d1c2ea871cddc78a51189e3ad2f18f63bd3d91" already present on machine kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:10:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsuitablecorgisimpleprod-1-fc8ddh9kr9.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:10:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsuitablecorgisimpleprod-1-fc8ddh9kr9.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:10:52 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsuitablecorgisimpleprod-1-fc8ddh9kr9.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:10:57 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsuitablecorgisimpleprod-1-fc8ddh9kr9.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:08 +0000 UTC Normal Pod simple-prod-collector-585c884f8c-ftrtp Binding Scheduled Successfully assigned kuttl-test-suitable-corgi/simple-prod-collector-585c884f8c-ftrtp to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:08 +0000 UTC Warning Pod simple-prod-collector-585c884f8c-ftrtp FailedMount MountVolume.SetUp failed for volume "simple-prod-collector-tls-config-volume" : secret "simple-prod-collector-headless-tls" not found kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:08 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-585c884f8c SuccessfulCreate Created pod: simple-prod-collector-585c884f8c-ftrtp replicaset-controller logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:08 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-585c884f8c to 1 deployment-controller logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:08 +0000 UTC Normal Pod simple-prod-query-56f74cc7f7-7hm6v Binding Scheduled Successfully assigned kuttl-test-suitable-corgi/simple-prod-query-56f74cc7f7-7hm6v to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:08 +0000 UTC Warning Pod simple-prod-query-56f74cc7f7-7hm6v FailedMount MountVolume.SetUp failed for volume "simple-prod-ui-oauth-proxy-tls" : secret "simple-prod-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:08 +0000 UTC Normal ReplicaSet.apps simple-prod-query-56f74cc7f7 SuccessfulCreate Created pod: simple-prod-query-56f74cc7f7-7hm6v replicaset-controller logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:08 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-56f74cc7f7 to 1 deployment-controller logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:09 +0000 UTC Normal Pod simple-prod-collector-585c884f8c-ftrtp AddedInterface Add eth0 [10.128.2.70/23] from ovn-kubernetes logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:09 +0000 UTC Normal Pod simple-prod-collector-585c884f8c-ftrtp.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:09 +0000 UTC Normal Pod simple-prod-collector-585c884f8c-ftrtp.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:09 +0000 UTC Normal Pod simple-prod-collector-585c884f8c-ftrtp.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:09 +0000 UTC Normal Pod simple-prod-query-56f74cc7f7-7hm6v AddedInterface Add eth0 [10.129.2.38/23] from ovn-kubernetes logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:09 +0000 UTC Normal Pod simple-prod-query-56f74cc7f7-7hm6v.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:09 +0000 UTC Normal Pod simple-prod-query-56f74cc7f7-7hm6v.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:09 +0000 UTC Normal Pod simple-prod-query-56f74cc7f7-7hm6v.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:09 +0000 UTC Normal Pod simple-prod-query-56f74cc7f7-7hm6v.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:09 +0000 UTC Normal Pod simple-prod-query-56f74cc7f7-7hm6v.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:09 +0000 UTC Normal Pod simple-prod-query-56f74cc7f7-7hm6v.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:09 +0000 UTC Normal Pod simple-prod-query-56f74cc7f7-7hm6v.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:09 +0000 UTC Normal Pod simple-prod-query-56f74cc7f7-7hm6v.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:09 +0000 UTC Normal Pod simple-prod-query-56f74cc7f7-7hm6v.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:14 +0000 UTC Normal Pod simple-prod-query-56f74cc7f7-7hm6v.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:14 +0000 UTC Normal ReplicaSet.apps simple-prod-query-56f74cc7f7 SuccessfulDelete Deleted pod: simple-prod-query-56f74cc7f7-7hm6v replicaset-controller logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:14 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-56f74cc7f7 to 0 from 1 deployment-controller logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:15 +0000 UTC Normal Pod simple-prod-query-56f74cc7f7-7hm6v.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:15 +0000 UTC Normal Pod simple-prod-query-56f74cc7f7-7hm6v.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:15 +0000 UTC Normal Pod simple-prod-query-cb6749595-jkvn9 Binding Scheduled Successfully assigned kuttl-test-suitable-corgi/simple-prod-query-cb6749595-jkvn9 to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:15 +0000 UTC Normal ReplicaSet.apps simple-prod-query-cb6749595 SuccessfulCreate Created pod: simple-prod-query-cb6749595-jkvn9 replicaset-controller logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:15 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-cb6749595 to 1 deployment-controller logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:16 +0000 UTC Normal Pod simple-prod-query-cb6749595-jkvn9 AddedInterface Add eth0 [10.129.2.39/23] from ovn-kubernetes logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:16 +0000 UTC Normal Pod simple-prod-query-cb6749595-jkvn9.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:16 +0000 UTC Normal Pod simple-prod-query-cb6749595-jkvn9.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:16 +0000 UTC Normal Pod simple-prod-query-cb6749595-jkvn9.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:16 +0000 UTC Normal Pod simple-prod-query-cb6749595-jkvn9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:16 +0000 UTC Normal Pod simple-prod-query-cb6749595-jkvn9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:16 +0000 UTC Normal Pod simple-prod-query-cb6749595-jkvn9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:16 +0000 UTC Normal Pod simple-prod-query-cb6749595-jkvn9.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:16 +0000 UTC Normal Pod simple-prod-query-cb6749595-jkvn9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:16 +0000 UTC Normal Pod simple-prod-query-cb6749595-jkvn9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:20 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-kwnld job-controller logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:20 +0000 UTC Normal Pod report-span-ggpqr Binding Scheduled Successfully assigned kuttl-test-suitable-corgi/report-span-ggpqr to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:20 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-ggpqr job-controller logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:21 +0000 UTC Normal Pod check-span-kwnld Binding Scheduled Successfully assigned kuttl-test-suitable-corgi/check-span-kwnld to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:21 +0000 UTC Normal Pod check-span-kwnld AddedInterface Add eth0 [10.128.2.72/23] from ovn-kubernetes logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:21 +0000 UTC Normal Pod check-span-kwnld.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:21 +0000 UTC Normal Pod check-span-kwnld.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:21 +0000 UTC Normal Pod check-span-kwnld.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:21 +0000 UTC Normal Pod report-span-ggpqr AddedInterface Add eth0 [10.128.2.71/23] from ovn-kubernetes logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:21 +0000 UTC Normal Pod report-span-ggpqr.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:21 +0000 UTC Normal Pod report-span-ggpqr.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:21 +0000 UTC Normal Pod report-span-ggpqr.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:23 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:23 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:23 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | 2024-01-22 07:11:31 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:11:32 | examples-simple-prod-with-volumes | Deleting namespace: kuttl-test-suitable-corgi === CONT kuttl/harness/examples-business-application-injected-sidecar logger.go:42: 07:11:51 | examples-business-application-injected-sidecar | Creating namespace: kuttl-test-light-gator logger.go:42: 07:11:51 | examples-business-application-injected-sidecar/0-install | starting test step 0-install logger.go:42: 07:11:52 | examples-business-application-injected-sidecar/0-install | Deployment:kuttl-test-light-gator/myapp created logger.go:42: 07:11:52 | examples-business-application-injected-sidecar/0-install | test step completed 0-install logger.go:42: 07:11:52 | examples-business-application-injected-sidecar/1-install | starting test step 1-install logger.go:42: 07:11:52 | examples-business-application-injected-sidecar/1-install | Jaeger:kuttl-test-light-gator/simplest created logger.go:42: 07:12:04 | examples-business-application-injected-sidecar/1-install | test step completed 1-install logger.go:42: 07:12:04 | examples-business-application-injected-sidecar/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:12:04 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:12:05 | examples-business-application-injected-sidecar/2-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:12:11 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:12:12 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:12:12 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/report-span created logger.go:42: 07:12:12 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/check-span created logger.go:42: 07:12:24 | examples-business-application-injected-sidecar/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | examples-business-application-injected-sidecar events from ns kuttl-test-light-gator: logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:11:52 +0000 UTC Normal Pod myapp-679f79d5f8-s4zjn Binding Scheduled Successfully assigned kuttl-test-light-gator/myapp-679f79d5f8-s4zjn to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:11:52 +0000 UTC Normal Pod myapp-679f79d5f8-s4zjn AddedInterface Add eth0 [10.131.0.44/23] from ovn-kubernetes logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:11:52 +0000 UTC Normal Pod myapp-679f79d5f8-s4zjn.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:11:52 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulCreate Created pod: myapp-679f79d5f8-s4zjn replicaset-controller logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:11:52 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-679f79d5f8 to 1 deployment-controller logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:11:55 +0000 UTC Normal Pod myapp-7c9cb8845f-88hxh Binding Scheduled Successfully assigned kuttl-test-light-gator/myapp-7c9cb8845f-88hxh to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:11:55 +0000 UTC Warning Pod myapp-7c9cb8845f-88hxh FailedMount MountVolume.SetUp failed for volume "simplest-service-ca" : configmap "simplest-service-ca" not found kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:11:55 +0000 UTC Warning Pod myapp-7c9cb8845f-88hxh FailedMount MountVolume.SetUp failed for volume "simplest-trusted-ca" : configmap "simplest-trusted-ca" not found kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:11:55 +0000 UTC Normal ReplicaSet.apps myapp-7c9cb8845f SuccessfulCreate Created pod: myapp-7c9cb8845f-88hxh replicaset-controller logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:11:55 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-7c9cb8845f to 1 deployment-controller logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:11:56 +0000 UTC Normal Pod myapp-679f79d5f8-s4zjn.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.975s (3.975s including waiting) kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:11:56 +0000 UTC Normal Pod myapp-679f79d5f8-s4zjn.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:11:56 +0000 UTC Normal Pod myapp-679f79d5f8-s4zjn.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:00 +0000 UTC Normal Pod simplest-89db57fcc-x2r56 Binding Scheduled Successfully assigned kuttl-test-light-gator/simplest-89db57fcc-x2r56 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:00 +0000 UTC Warning Pod simplest-89db57fcc-x2r56 FailedMount MountVolume.SetUp failed for volume "simplest-collector-tls-config-volume" : secret "simplest-collector-headless-tls" not found kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:00 +0000 UTC Normal ReplicaSet.apps simplest-89db57fcc SuccessfulCreate Created pod: simplest-89db57fcc-x2r56 replicaset-controller logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:00 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-89db57fcc to 1 deployment-controller logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:01 +0000 UTC Warning Pod simplest-89db57fcc-x2r56 FailedMount MountVolume.SetUp failed for volume "simplest-ui-oauth-proxy-tls" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:02 +0000 UTC Normal Pod simplest-89db57fcc-x2r56 AddedInterface Add eth0 [10.131.0.45/23] from ovn-kubernetes logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:02 +0000 UTC Normal Pod simplest-89db57fcc-x2r56.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:02 +0000 UTC Normal Pod simplest-89db57fcc-x2r56.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:02 +0000 UTC Normal Pod simplest-89db57fcc-x2r56.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:02 +0000 UTC Normal Pod simplest-89db57fcc-x2r56.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:02 +0000 UTC Normal Pod simplest-89db57fcc-x2r56.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:02 +0000 UTC Normal Pod simplest-89db57fcc-x2r56.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:03 +0000 UTC Warning Pod myapp-679f79d5f8-s4zjn.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.131.0.44:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:03 +0000 UTC Normal Pod myapp-7c9cb8845f-88hxh AddedInterface Add eth0 [10.129.2.40/23] from ovn-kubernetes logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:03 +0000 UTC Normal Pod myapp-7c9cb8845f-88hxh.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:06 +0000 UTC Normal ReplicaSet.apps simplest-89db57fcc SuccessfulDelete Deleted pod: simplest-89db57fcc-x2r56 replicaset-controller logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:06 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-89db57fcc to 0 from 1 deployment-controller logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:07 +0000 UTC Normal Pod myapp-7c9cb8845f-88hxh.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.726s (3.726s including waiting) kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:07 +0000 UTC Normal Pod myapp-7c9cb8845f-88hxh.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:07 +0000 UTC Normal Pod myapp-7c9cb8845f-88hxh.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:07 +0000 UTC Normal Pod myapp-7c9cb8845f-88hxh.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:07 +0000 UTC Normal Pod myapp-7c9cb8845f-88hxh.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:07 +0000 UTC Normal Pod myapp-7c9cb8845f-88hxh.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:07 +0000 UTC Normal Pod simplest-89db57fcc-x2r56.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:07 +0000 UTC Normal Pod simplest-89db57fcc-x2r56.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:08 +0000 UTC Normal Pod myapp-679f79d5f8-s4zjn.spec.containers{myapp} Killing Stopping container myapp kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:08 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulDelete Deleted pod: myapp-679f79d5f8-s4zjn replicaset-controller logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:08 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled down replica set myapp-679f79d5f8 to 0 from 1 deployment-controller logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:08 +0000 UTC Normal Pod simplest-696f77f6b8-m8lm6 Binding Scheduled Successfully assigned kuttl-test-light-gator/simplest-696f77f6b8-m8lm6 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:08 +0000 UTC Normal Pod simplest-696f77f6b8-m8lm6 AddedInterface Add eth0 [10.131.0.46/23] from ovn-kubernetes logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:08 +0000 UTC Normal Pod simplest-696f77f6b8-m8lm6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:08 +0000 UTC Normal Pod simplest-696f77f6b8-m8lm6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:08 +0000 UTC Normal Pod simplest-696f77f6b8-m8lm6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:08 +0000 UTC Normal Pod simplest-696f77f6b8-m8lm6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:08 +0000 UTC Normal Pod simplest-696f77f6b8-m8lm6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:08 +0000 UTC Normal Pod simplest-696f77f6b8-m8lm6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:08 +0000 UTC Normal ReplicaSet.apps simplest-696f77f6b8 SuccessfulCreate Created pod: simplest-696f77f6b8-m8lm6 replicaset-controller logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:08 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-696f77f6b8 to 1 deployment-controller logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:12 +0000 UTC Normal Pod check-span-9w8m5 Binding Scheduled Successfully assigned kuttl-test-light-gator/check-span-9w8m5 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:12 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-9w8m5 job-controller logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:12 +0000 UTC Normal Pod report-span-4zzjz Binding Scheduled Successfully assigned kuttl-test-light-gator/report-span-4zzjz to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:12 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-4zzjz job-controller logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:13 +0000 UTC Normal Pod check-span-9w8m5 AddedInterface Add eth0 [10.128.2.74/23] from ovn-kubernetes logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:13 +0000 UTC Normal Pod check-span-9w8m5.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:13 +0000 UTC Normal Pod check-span-9w8m5.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:13 +0000 UTC Normal Pod check-span-9w8m5.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:13 +0000 UTC Normal Pod report-span-4zzjz AddedInterface Add eth0 [10.128.2.73/23] from ovn-kubernetes logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:13 +0000 UTC Normal Pod report-span-4zzjz.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:13 +0000 UTC Normal Pod report-span-4zzjz.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:13 +0000 UTC Normal Pod report-span-4zzjz.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:14 +0000 UTC Warning Pod myapp-7c9cb8845f-88hxh.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.129.2.40:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | 2024-01-22 07:12:24 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:12:24 | examples-business-application-injected-sidecar | Deleting namespace: kuttl-test-light-gator === CONT kuttl/harness/examples-service-types logger.go:42: 07:12:32 | examples-service-types | Creating namespace: kuttl-test-special-newt logger.go:42: 07:12:32 | examples-service-types/0-install | starting test step 0-install logger.go:42: 07:12:32 | examples-service-types/0-install | Jaeger:kuttl-test-special-newt/service-types created logger.go:42: 07:12:38 | examples-service-types/0-install | test step completed 0-install logger.go:42: 07:12:38 | examples-service-types/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:12:38 | examples-service-types/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE service-types /dev/null] logger.go:42: 07:12:39 | examples-service-types/1-smoke-test | Warning: resource jaegers/service-types is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:12:46 | examples-service-types/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://service-types-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:12:46 | examples-service-types/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:12:47 | examples-service-types/1-smoke-test | job.batch/report-span created logger.go:42: 07:12:47 | examples-service-types/1-smoke-test | job.batch/check-span created logger.go:42: 07:12:58 | examples-service-types/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:12:58 | examples-service-types/2- | starting test step 2- logger.go:42: 07:12:58 | examples-service-types/2- | test step completed 2- logger.go:42: 07:12:58 | examples-service-types | examples-service-types events from ns kuttl-test-special-newt: logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:35 +0000 UTC Normal Pod service-types-7764d759b6-fhc55 Binding Scheduled Successfully assigned kuttl-test-special-newt/service-types-7764d759b6-fhc55 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:35 +0000 UTC Normal ReplicaSet.apps service-types-7764d759b6 SuccessfulCreate Created pod: service-types-7764d759b6-fhc55 replicaset-controller logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:35 +0000 UTC Normal Service service-types-collector EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:35 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-7764d759b6 to 1 deployment-controller logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:36 +0000 UTC Warning Pod service-types-7764d759b6-fhc55 FailedMount MountVolume.SetUp failed for volume "service-types-ui-oauth-proxy-tls" : secret "service-types-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:37 +0000 UTC Normal Pod service-types-7764d759b6-fhc55 AddedInterface Add eth0 [10.131.0.47/23] from ovn-kubernetes logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:37 +0000 UTC Normal Pod service-types-7764d759b6-fhc55.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:37 +0000 UTC Normal Pod service-types-7764d759b6-fhc55.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:37 +0000 UTC Normal Pod service-types-7764d759b6-fhc55.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:37 +0000 UTC Normal Pod service-types-7764d759b6-fhc55.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:37 +0000 UTC Normal Pod service-types-7764d759b6-fhc55.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:37 +0000 UTC Normal Pod service-types-7764d759b6-fhc55.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:39 +0000 UTC Normal Service service-types-collector EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:39 +0000 UTC Normal Service service-types-query EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:42 +0000 UTC Normal Pod service-types-7764d759b6-fhc55.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:42 +0000 UTC Normal Pod service-types-7764d759b6-fhc55.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:42 +0000 UTC Normal ReplicaSet.apps service-types-7764d759b6 SuccessfulDelete Deleted pod: service-types-7764d759b6-fhc55 replicaset-controller logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:42 +0000 UTC Normal Service service-types-query EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:42 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled down replica set service-types-7764d759b6 to 0 from 1 deployment-controller logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:43 +0000 UTC Normal Pod service-types-557696787-dxsrh Binding Scheduled Successfully assigned kuttl-test-special-newt/service-types-557696787-dxsrh to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:43 +0000 UTC Normal Pod service-types-557696787-dxsrh AddedInterface Add eth0 [10.131.0.48/23] from ovn-kubernetes logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:43 +0000 UTC Normal Pod service-types-557696787-dxsrh.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:43 +0000 UTC Normal Pod service-types-557696787-dxsrh.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:43 +0000 UTC Normal Pod service-types-557696787-dxsrh.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:43 +0000 UTC Normal ReplicaSet.apps service-types-557696787 SuccessfulCreate Created pod: service-types-557696787-dxsrh replicaset-controller logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:43 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-557696787 to 1 deployment-controller logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:44 +0000 UTC Normal Pod service-types-557696787-dxsrh.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:44 +0000 UTC Normal Pod service-types-557696787-dxsrh.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:44 +0000 UTC Normal Pod service-types-557696787-dxsrh.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:47 +0000 UTC Normal Pod check-span-dvnkv Binding Scheduled Successfully assigned kuttl-test-special-newt/check-span-dvnkv to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:47 +0000 UTC Normal Pod check-span-dvnkv AddedInterface Add eth0 [10.129.2.41/23] from ovn-kubernetes logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:47 +0000 UTC Normal Pod check-span-dvnkv.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:47 +0000 UTC Normal Pod check-span-dvnkv.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:47 +0000 UTC Normal Pod check-span-dvnkv.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:47 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-dvnkv job-controller logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:47 +0000 UTC Normal Pod report-span-rlvxf Binding Scheduled Successfully assigned kuttl-test-special-newt/report-span-rlvxf to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:47 +0000 UTC Normal Pod report-span-rlvxf AddedInterface Add eth0 [10.128.2.75/23] from ovn-kubernetes logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:47 +0000 UTC Normal Pod report-span-rlvxf.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:47 +0000 UTC Normal Pod report-span-rlvxf.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:47 +0000 UTC Normal Pod report-span-rlvxf.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:47 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-rlvxf job-controller logger.go:42: 07:12:58 | examples-service-types | 2024-01-22 07:12:58 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:12:58 | examples-service-types | Deleting namespace: kuttl-test-special-newt === CONT kuttl/harness/examples-openshift-with-htpasswd logger.go:42: 07:13:30 | examples-openshift-with-htpasswd | Ignoring 00-install.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:13:30 | examples-openshift-with-htpasswd | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:13:30 | examples-openshift-with-htpasswd | Creating namespace: kuttl-test-rational-boar logger.go:42: 07:13:30 | examples-openshift-with-htpasswd/0-install | starting test step 0-install logger.go:42: 07:13:30 | examples-openshift-with-htpasswd/0-install | Secret:kuttl-test-rational-boar/htpasswd created logger.go:42: 07:13:30 | examples-openshift-with-htpasswd/0-install | test step completed 0-install logger.go:42: 07:13:30 | examples-openshift-with-htpasswd/1-install | starting test step 1-install logger.go:42: 07:13:30 | examples-openshift-with-htpasswd/1-install | Jaeger:kuttl-test-rational-boar/with-htpasswd created logger.go:42: 07:13:38 | examples-openshift-with-htpasswd/1-install | test step completed 1-install logger.go:42: 07:13:38 | examples-openshift-with-htpasswd/2-check-unsecured | starting test step 2-check-unsecured logger.go:42: 07:13:38 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [./ensure-ingress-host.sh] logger.go:42: 07:13:38 | examples-openshift-with-htpasswd/2-check-unsecured | Checking the Ingress host value was populated logger.go:42: 07:13:38 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 0 logger.go:42: 07:13:38 | examples-openshift-with-htpasswd/2-check-unsecured | Hostname is with-htpasswd-kuttl-test-rational-boar.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:13:38 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:13:38 | examples-openshift-with-htpasswd/2-check-unsecured | Checking an expected HTTP response logger.go:42: 07:13:38 | examples-openshift-with-htpasswd/2-check-unsecured | Running in OpenShift logger.go:42: 07:13:38 | examples-openshift-with-htpasswd/2-check-unsecured | Not using any secret logger.go:42: 07:13:38 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 1/30 the https://with-htpasswd-kuttl-test-rational-boar.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:13:38 | examples-openshift-with-htpasswd/2-check-unsecured | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:13:38 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 2/30 the https://with-htpasswd-kuttl-test-rational-boar.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:13:38 | examples-openshift-with-htpasswd/2-check-unsecured | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 3/30 the https://with-htpasswd-kuttl-test-rational-boar.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/2-check-unsecured | curl response asserted properly logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/2-check-unsecured | test step completed 2-check-unsecured logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/3-check-unauthorized | starting test step 3-check-unauthorized logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking the Ingress host value was populated logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 0 logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/3-check-unauthorized | Hostname is with-htpasswd-kuttl-test-rational-boar.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [sh -c JAEGER_USERNAME=wronguser JAEGER_PASSWORD=wrongpassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking an expected HTTP response logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/3-check-unauthorized | Running in OpenShift logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/3-check-unauthorized | Using Jaeger basic authentication logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 1/30 the https://with-htpasswd-kuttl-test-rational-boar.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/3-check-unauthorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 2/30 the https://with-htpasswd-kuttl-test-rational-boar.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/3-check-unauthorized | curl response asserted properly logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/3-check-unauthorized | test step completed 3-check-unauthorized logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/4-check-authorized | starting test step 4-check-authorized logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/4-check-authorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/4-check-authorized | Checking the Ingress host value was populated logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/4-check-authorized | Try number 0 logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/4-check-authorized | Hostname is with-htpasswd-kuttl-test-rational-boar.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/4-check-authorized | running command: [sh -c JAEGER_USERNAME=awesomeuser JAEGER_PASSWORD=awesomepassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE with-htpasswd] logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/4-check-authorized | Checking an expected HTTP response logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/4-check-authorized | Running in OpenShift logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/4-check-authorized | Using Jaeger basic authentication logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/4-check-authorized | Try number 1/30 the https://with-htpasswd-kuttl-test-rational-boar.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/4-check-authorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/4-check-authorized | Try number 2/30 the https://with-htpasswd-kuttl-test-rational-boar.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/4-check-authorized | curl response asserted properly logger.go:42: 07:13:48 | examples-openshift-with-htpasswd/4-check-authorized | test step completed 4-check-authorized logger.go:42: 07:13:48 | examples-openshift-with-htpasswd | examples-openshift-with-htpasswd events from ns kuttl-test-rational-boar: logger.go:42: 07:13:48 | examples-openshift-with-htpasswd | 2024-01-22 07:13:34 +0000 UTC Normal Pod with-htpasswd-bcf698474-9dsbs Binding Scheduled Successfully assigned kuttl-test-rational-boar/with-htpasswd-bcf698474-9dsbs to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:13:48 | examples-openshift-with-htpasswd | 2024-01-22 07:13:34 +0000 UTC Normal ReplicaSet.apps with-htpasswd-bcf698474 SuccessfulCreate Created pod: with-htpasswd-bcf698474-9dsbs replicaset-controller logger.go:42: 07:13:48 | examples-openshift-with-htpasswd | 2024-01-22 07:13:34 +0000 UTC Normal Deployment.apps with-htpasswd ScalingReplicaSet Scaled up replica set with-htpasswd-bcf698474 to 1 deployment-controller logger.go:42: 07:13:48 | examples-openshift-with-htpasswd | 2024-01-22 07:13:35 +0000 UTC Normal Pod with-htpasswd-bcf698474-9dsbs AddedInterface Add eth0 [10.131.0.49/23] from ovn-kubernetes logger.go:42: 07:13:48 | examples-openshift-with-htpasswd | 2024-01-22 07:13:35 +0000 UTC Normal Pod with-htpasswd-bcf698474-9dsbs.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:13:48 | examples-openshift-with-htpasswd | 2024-01-22 07:13:35 +0000 UTC Normal Pod with-htpasswd-bcf698474-9dsbs.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:13:48 | examples-openshift-with-htpasswd | 2024-01-22 07:13:35 +0000 UTC Normal Pod with-htpasswd-bcf698474-9dsbs.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:13:48 | examples-openshift-with-htpasswd | 2024-01-22 07:13:35 +0000 UTC Normal Pod with-htpasswd-bcf698474-9dsbs.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:13:48 | examples-openshift-with-htpasswd | 2024-01-22 07:13:35 +0000 UTC Normal Pod with-htpasswd-bcf698474-9dsbs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:13:48 | examples-openshift-with-htpasswd | 2024-01-22 07:13:35 +0000 UTC Normal Pod with-htpasswd-bcf698474-9dsbs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:13:48 | examples-openshift-with-htpasswd | Deleting namespace: kuttl-test-rational-boar === CONT kuttl/harness/examples-collector-with-priority-class logger.go:42: 07:13:55 | examples-collector-with-priority-class | Creating namespace: kuttl-test-safe-oarfish logger.go:42: 07:13:55 | examples-collector-with-priority-class/0-install | starting test step 0-install logger.go:42: 07:13:55 | examples-collector-with-priority-class/0-install | PriorityClass:/collector-high-priority created logger.go:42: 07:13:55 | examples-collector-with-priority-class/0-install | Jaeger:kuttl-test-safe-oarfish/collector-with-high-priority created logger.go:42: 07:14:01 | examples-collector-with-priority-class/0-install | test step completed 0-install logger.go:42: 07:14:01 | examples-collector-with-priority-class/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:14:01 | examples-collector-with-priority-class/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE collector-with-high-priority /dev/null] logger.go:42: 07:14:02 | examples-collector-with-priority-class/1-smoke-test | Warning: resource jaegers/collector-with-high-priority is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:14:08 | examples-collector-with-priority-class/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:14:09 | examples-collector-with-priority-class/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:14:09 | examples-collector-with-priority-class/1-smoke-test | job.batch/report-span created logger.go:42: 07:14:09 | examples-collector-with-priority-class/1-smoke-test | job.batch/check-span created logger.go:42: 07:14:21 | examples-collector-with-priority-class/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:14:21 | examples-collector-with-priority-class | examples-collector-with-priority-class events from ns kuttl-test-safe-oarfish: logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:13:58 +0000 UTC Normal Pod collector-with-high-priority-77b97855cf-xqmwp Binding Scheduled Successfully assigned kuttl-test-safe-oarfish/collector-with-high-priority-77b97855cf-xqmwp to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:13:58 +0000 UTC Normal ReplicaSet.apps collector-with-high-priority-77b97855cf SuccessfulCreate Created pod: collector-with-high-priority-77b97855cf-xqmwp replicaset-controller logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:13:58 +0000 UTC Normal Deployment.apps collector-with-high-priority ScalingReplicaSet Scaled up replica set collector-with-high-priority-77b97855cf to 1 deployment-controller logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:13:59 +0000 UTC Normal Pod collector-with-high-priority-77b97855cf-xqmwp AddedInterface Add eth0 [10.131.0.50/23] from ovn-kubernetes logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:13:59 +0000 UTC Normal Pod collector-with-high-priority-77b97855cf-xqmwp.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:13:59 +0000 UTC Normal Pod collector-with-high-priority-77b97855cf-xqmwp.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:13:59 +0000 UTC Normal Pod collector-with-high-priority-77b97855cf-xqmwp.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:13:59 +0000 UTC Normal Pod collector-with-high-priority-77b97855cf-xqmwp.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:13:59 +0000 UTC Normal Pod collector-with-high-priority-77b97855cf-xqmwp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:13:59 +0000 UTC Normal Pod collector-with-high-priority-77b97855cf-xqmwp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:03 +0000 UTC Normal Pod collector-with-high-priority-77b97855cf-xqmwp.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:03 +0000 UTC Normal Pod collector-with-high-priority-77b97855cf-xqmwp.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:03 +0000 UTC Normal ReplicaSet.apps collector-with-high-priority-77b97855cf SuccessfulDelete Deleted pod: collector-with-high-priority-77b97855cf-xqmwp replicaset-controller logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:03 +0000 UTC Normal Deployment.apps collector-with-high-priority ScalingReplicaSet Scaled down replica set collector-with-high-priority-77b97855cf to 0 from 1 deployment-controller logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:04 +0000 UTC Normal Pod collector-with-high-priority-8d4dcd586-rnmf5 Binding Scheduled Successfully assigned kuttl-test-safe-oarfish/collector-with-high-priority-8d4dcd586-rnmf5 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:04 +0000 UTC Normal ReplicaSet.apps collector-with-high-priority-8d4dcd586 SuccessfulCreate Created pod: collector-with-high-priority-8d4dcd586-rnmf5 replicaset-controller logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:04 +0000 UTC Normal Deployment.apps collector-with-high-priority ScalingReplicaSet Scaled up replica set collector-with-high-priority-8d4dcd586 to 1 deployment-controller logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:05 +0000 UTC Normal Pod collector-with-high-priority-8d4dcd586-rnmf5 AddedInterface Add eth0 [10.131.0.51/23] from ovn-kubernetes logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:05 +0000 UTC Normal Pod collector-with-high-priority-8d4dcd586-rnmf5.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:05 +0000 UTC Normal Pod collector-with-high-priority-8d4dcd586-rnmf5.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:05 +0000 UTC Normal Pod collector-with-high-priority-8d4dcd586-rnmf5.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:05 +0000 UTC Normal Pod collector-with-high-priority-8d4dcd586-rnmf5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:05 +0000 UTC Normal Pod collector-with-high-priority-8d4dcd586-rnmf5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:05 +0000 UTC Normal Pod collector-with-high-priority-8d4dcd586-rnmf5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:09 +0000 UTC Normal Pod check-span-jk2nn Binding Scheduled Successfully assigned kuttl-test-safe-oarfish/check-span-jk2nn to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:09 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-jk2nn job-controller logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:09 +0000 UTC Normal Pod report-span-h5qpw Binding Scheduled Successfully assigned kuttl-test-safe-oarfish/report-span-h5qpw to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:09 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-h5qpw job-controller logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:10 +0000 UTC Normal Pod check-span-jk2nn AddedInterface Add eth0 [10.129.2.42/23] from ovn-kubernetes logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:10 +0000 UTC Normal Pod check-span-jk2nn.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:10 +0000 UTC Normal Pod check-span-jk2nn.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:10 +0000 UTC Normal Pod check-span-jk2nn.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:10 +0000 UTC Normal Pod report-span-h5qpw AddedInterface Add eth0 [10.128.2.76/23] from ovn-kubernetes logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:10 +0000 UTC Normal Pod report-span-h5qpw.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:10 +0000 UTC Normal Pod report-span-h5qpw.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:10 +0000 UTC Normal Pod report-span-h5qpw.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:14:21 | examples-collector-with-priority-class | 2024-01-22 07:14:21 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:14:21 | examples-collector-with-priority-class | Deleting namespace: kuttl-test-safe-oarfish === CONT kuttl/harness/examples-agent-with-priority-class logger.go:42: 07:14:33 | examples-agent-with-priority-class | Creating namespace: kuttl-test-pet-ant logger.go:42: 07:14:33 | examples-agent-with-priority-class/0-install | starting test step 0-install logger.go:42: 07:14:34 | examples-agent-with-priority-class/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 07:14:34 | examples-agent-with-priority-class/0-install | ServiceAccount:kuttl-test-pet-ant/jaeger-agent-daemonset created logger.go:42: 07:14:34 | examples-agent-with-priority-class/0-install | test step completed 0-install logger.go:42: 07:14:34 | examples-agent-with-priority-class/1-add-policy | starting test step 1-add-policy logger.go:42: 07:14:34 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 07:14:34 | examples-agent-with-priority-class/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 07:14:34 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 07:14:39 | examples-agent-with-priority-class/1-add-policy | test step completed 1-add-policy logger.go:42: 07:14:39 | examples-agent-with-priority-class/2-install | starting test step 2-install logger.go:42: 07:14:39 | examples-agent-with-priority-class/2-install | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 07:14:46 | examples-agent-with-priority-class/2-install | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:14:47 | examples-agent-with-priority-class/2-install | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:14:47 | examples-agent-with-priority-class/2-install | job.batch/report-span created logger.go:42: 07:14:47 | examples-agent-with-priority-class/2-install | job.batch/check-span created logger.go:42: 07:14:47 | examples-agent-with-priority-class/2-install | PriorityClass:/high-priority created logger.go:42: 07:14:47 | examples-agent-with-priority-class/2-install | Jaeger:kuttl-test-pet-ant/agent-as-daemonset updated logger.go:42: 07:14:58 | examples-agent-with-priority-class/2-install | test step completed 2-install logger.go:42: 07:14:58 | examples-agent-with-priority-class | examples-agent-with-priority-class events from ns kuttl-test-pet-ant: logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:44 +0000 UTC Normal Pod agent-as-daemonset-845df98bf4-rr2b9 Binding Scheduled Successfully assigned kuttl-test-pet-ant/agent-as-daemonset-845df98bf4-rr2b9 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:44 +0000 UTC Warning Pod agent-as-daemonset-845df98bf4-rr2b9 FailedMount MountVolume.SetUp failed for volume "agent-as-daemonset-ui-oauth-proxy-tls" : secret "agent-as-daemonset-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:44 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-845df98bf4 SuccessfulCreate Created pod: agent-as-daemonset-845df98bf4-rr2b9 replicaset-controller logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:44 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-845df98bf4 to 1 deployment-controller logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:45 +0000 UTC Normal Pod agent-as-daemonset-845df98bf4-rr2b9 AddedInterface Add eth0 [10.131.0.52/23] from ovn-kubernetes logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:45 +0000 UTC Normal Pod agent-as-daemonset-845df98bf4-rr2b9.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:45 +0000 UTC Normal Pod agent-as-daemonset-845df98bf4-rr2b9.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:45 +0000 UTC Normal Pod agent-as-daemonset-845df98bf4-rr2b9.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:45 +0000 UTC Normal Pod agent-as-daemonset-845df98bf4-rr2b9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:45 +0000 UTC Normal Pod agent-as-daemonset-845df98bf4-rr2b9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:45 +0000 UTC Normal Pod agent-as-daemonset-845df98bf4-rr2b9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:47 +0000 UTC Normal Pod check-span-c6brr Binding Scheduled Successfully assigned kuttl-test-pet-ant/check-span-c6brr to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:47 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-c6brr job-controller logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:47 +0000 UTC Normal Pod report-span-j4vk5 Binding Scheduled Successfully assigned kuttl-test-pet-ant/report-span-j4vk5 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:47 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-j4vk5 job-controller logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:48 +0000 UTC Warning DaemonSet.apps agent-as-daemonset-agent-daemonset FailedCreate Error creating: pods "agent-as-daemonset-agent-daemonset-" is forbidden: unable to validate against any security context constraint: [provider "anyuid": Forbidden: not usable by user or serviceaccount, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5775: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5778: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6831: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6832: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 14271: Host ports are not allowed to be used, provider "restricted": Forbidden: not usable by user or serviceaccount, provider "nonroot-v2": Forbidden: not usable by user or serviceaccount, provider "nonroot": Forbidden: not usable by user or serviceaccount, provider "hostmount-anyuid": Forbidden: not usable by user or serviceaccount, provider "elasticsearch-scc": Forbidden: not usable by user or serviceaccount, provider "machine-api-termination-handler": Forbidden: not usable by user or serviceaccount, provider "daemonset-with-hostport": Forbidden: not usable by user or serviceaccount, provider "hostnetwork-v2": Forbidden: not usable by user or serviceaccount, provider "hostnetwork": Forbidden: not usable by user or serviceaccount, provider "hostaccess": Forbidden: not usable by user or serviceaccount, provider "node-exporter": Forbidden: not usable by user or serviceaccount, provider "privileged": Forbidden: not usable by user or serviceaccount] daemonset-controller logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:48 +0000 UTC Normal Pod check-span-c6brr AddedInterface Add eth0 [10.129.2.43/23] from ovn-kubernetes logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:48 +0000 UTC Normal Pod check-span-c6brr.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:48 +0000 UTC Normal Pod check-span-c6brr.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:48 +0000 UTC Normal Pod check-span-c6brr.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:48 +0000 UTC Normal Pod report-span-j4vk5 AddedInterface Add eth0 [10.128.2.77/23] from ovn-kubernetes logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:48 +0000 UTC Normal Pod report-span-j4vk5.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:48 +0000 UTC Normal Pod report-span-j4vk5.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:48 +0000 UTC Normal Pod report-span-j4vk5.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:14:58 | examples-agent-with-priority-class | 2024-01-22 07:14:58 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:14:58 | examples-agent-with-priority-class | Deleting namespace: kuttl-test-pet-ant === CONT kuttl/harness/examples-all-in-one-with-options logger.go:42: 07:15:27 | examples-all-in-one-with-options | Creating namespace: kuttl-test-immense-mudfish logger.go:42: 07:15:27 | examples-all-in-one-with-options/0-install | starting test step 0-install logger.go:42: 07:15:27 | examples-all-in-one-with-options/0-install | Jaeger:kuttl-test-immense-mudfish/my-jaeger created logger.go:42: 07:15:34 | examples-all-in-one-with-options/0-install | test step completed 0-install logger.go:42: 07:15:34 | examples-all-in-one-with-options/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:15:34 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:15:35 | examples-all-in-one-with-options/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:15:42 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443/jaeger MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:15:42 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:15:42 | examples-all-in-one-with-options/1-smoke-test | job.batch/report-span created logger.go:42: 07:15:42 | examples-all-in-one-with-options/1-smoke-test | job.batch/check-span created logger.go:42: 07:15:54 | examples-all-in-one-with-options/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:15:54 | examples-all-in-one-with-options | examples-all-in-one-with-options events from ns kuttl-test-immense-mudfish: logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:31 +0000 UTC Normal Pod my-jaeger-7c57f88f5d-kv2mc Binding Scheduled Successfully assigned kuttl-test-immense-mudfish/my-jaeger-7c57f88f5d-kv2mc to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:31 +0000 UTC Warning Pod my-jaeger-7c57f88f5d-kv2mc FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:31 +0000 UTC Warning Pod my-jaeger-7c57f88f5d-kv2mc FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-oauth-proxy-tls" : secret "my-jaeger-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:31 +0000 UTC Normal ReplicaSet.apps my-jaeger-7c57f88f5d SuccessfulCreate Created pod: my-jaeger-7c57f88f5d-kv2mc replicaset-controller logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:31 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-7c57f88f5d to 1 deployment-controller logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:32 +0000 UTC Normal Pod my-jaeger-7c57f88f5d-kv2mc AddedInterface Add eth0 [10.131.0.53/23] from ovn-kubernetes logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:32 +0000 UTC Normal Pod my-jaeger-7c57f88f5d-kv2mc.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:32 +0000 UTC Normal Pod my-jaeger-7c57f88f5d-kv2mc.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:32 +0000 UTC Normal Pod my-jaeger-7c57f88f5d-kv2mc.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:32 +0000 UTC Normal Pod my-jaeger-7c57f88f5d-kv2mc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:32 +0000 UTC Normal Pod my-jaeger-7c57f88f5d-kv2mc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:33 +0000 UTC Normal Pod my-jaeger-7c57f88f5d-kv2mc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:37 +0000 UTC Normal Pod my-jaeger-7c57f88f5d-kv2mc.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:37 +0000 UTC Normal Pod my-jaeger-7c57f88f5d-kv2mc.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:37 +0000 UTC Normal ReplicaSet.apps my-jaeger-7c57f88f5d SuccessfulDelete Deleted pod: my-jaeger-7c57f88f5d-kv2mc replicaset-controller logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:37 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-7c57f88f5d to 0 from 1 deployment-controller logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:38 +0000 UTC Normal Pod my-jaeger-6cb7dd944d-lll9n Binding Scheduled Successfully assigned kuttl-test-immense-mudfish/my-jaeger-6cb7dd944d-lll9n to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:38 +0000 UTC Normal ReplicaSet.apps my-jaeger-6cb7dd944d SuccessfulCreate Created pod: my-jaeger-6cb7dd944d-lll9n replicaset-controller logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:38 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6cb7dd944d to 1 deployment-controller logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:39 +0000 UTC Normal Pod my-jaeger-6cb7dd944d-lll9n AddedInterface Add eth0 [10.131.0.54/23] from ovn-kubernetes logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:39 +0000 UTC Normal Pod my-jaeger-6cb7dd944d-lll9n.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:39 +0000 UTC Normal Pod my-jaeger-6cb7dd944d-lll9n.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:39 +0000 UTC Normal Pod my-jaeger-6cb7dd944d-lll9n.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:39 +0000 UTC Normal Pod my-jaeger-6cb7dd944d-lll9n.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:39 +0000 UTC Normal Pod my-jaeger-6cb7dd944d-lll9n.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:39 +0000 UTC Normal Pod my-jaeger-6cb7dd944d-lll9n.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:42 +0000 UTC Normal Pod check-span-j9c5t Binding Scheduled Successfully assigned kuttl-test-immense-mudfish/check-span-j9c5t to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:42 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-j9c5t job-controller logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:42 +0000 UTC Normal Pod report-span-7cq4d Binding Scheduled Successfully assigned kuttl-test-immense-mudfish/report-span-7cq4d to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:42 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-7cq4d job-controller logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:43 +0000 UTC Normal Pod check-span-j9c5t AddedInterface Add eth0 [10.129.2.44/23] from ovn-kubernetes logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:43 +0000 UTC Normal Pod check-span-j9c5t.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:43 +0000 UTC Normal Pod check-span-j9c5t.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:43 +0000 UTC Normal Pod check-span-j9c5t.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:43 +0000 UTC Normal Pod report-span-7cq4d AddedInterface Add eth0 [10.128.2.79/23] from ovn-kubernetes logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:43 +0000 UTC Normal Pod report-span-7cq4d.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:43 +0000 UTC Normal Pod report-span-7cq4d.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:43 +0000 UTC Normal Pod report-span-7cq4d.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:15:54 | examples-all-in-one-with-options | 2024-01-22 07:15:54 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:15:54 | examples-all-in-one-with-options | Deleting namespace: kuttl-test-immense-mudfish === CONT kuttl/harness/examples-agent-as-daemonset logger.go:42: 07:16:01 | examples-agent-as-daemonset | Creating namespace: kuttl-test-loved-goldfish logger.go:42: 07:16:01 | examples-agent-as-daemonset/0-install | starting test step 0-install logger.go:42: 07:16:01 | examples-agent-as-daemonset/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 07:16:01 | examples-agent-as-daemonset/0-install | ServiceAccount:kuttl-test-loved-goldfish/jaeger-agent-daemonset created logger.go:42: 07:16:01 | examples-agent-as-daemonset/0-install | test step completed 0-install logger.go:42: 07:16:01 | examples-agent-as-daemonset/1-add-policy | starting test step 1-add-policy logger.go:42: 07:16:01 | examples-agent-as-daemonset/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 07:16:01 | examples-agent-as-daemonset/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 07:16:01 | examples-agent-as-daemonset/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 07:16:06 | examples-agent-as-daemonset/1-add-policy | test step completed 1-add-policy logger.go:42: 07:16:06 | examples-agent-as-daemonset/2-install | starting test step 2-install logger.go:42: 07:16:06 | examples-agent-as-daemonset/2-install | Jaeger:kuttl-test-loved-goldfish/agent-as-daemonset created logger.go:42: 07:16:06 | examples-agent-as-daemonset/2-install | test step completed 2-install logger.go:42: 07:16:06 | examples-agent-as-daemonset/3- | starting test step 3- logger.go:42: 07:26:07 | examples-agent-as-daemonset/3- | test step failed 3- case.go:364: failed in step 3- case.go:366: --- DaemonSet:kuttl-test-loved-goldfish/agent-as-daemonset-agent-daemonset +++ DaemonSet:kuttl-test-loved-goldfish/agent-as-daemonset-agent-daemonset @@ -1,8 +1,310 @@ apiVersion: apps/v1 kind: DaemonSet metadata: + annotations: + deprecated.daemonset.template.generation: "1" + labels: + app: jaeger + app.kubernetes.io/component: agent + app.kubernetes.io/instance: agent-as-daemonset + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: agent-as-daemonset-agent + app.kubernetes.io/part-of: jaeger + managedFields: + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:deprecated.daemonset.template.generation: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:ownerReferences: + .: {} + k:{"uid":"6fe15c79-36d0-477d-861d-ee90ef391c16"}: {} + f:spec: + f:revisionHistoryLimit: {} + f:selector: {} + f:template: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.istio.io/inject: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:spec: + f:containers: + k:{"name":"jaeger-agent-daemonset"}: + .: {} + f:args: {} + f:image: {} + f:imagePullPolicy: {} + f:livenessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":5775,"protocol":"UDP"}: + .: {} + f:containerPort: {} + f:hostPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":5778,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:hostPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":6831,"protocol":"UDP"}: + .: {} + f:containerPort: {} + f:hostPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":6832,"protocol":"UDP"}: + .: {} + f:containerPort: {} + f:hostPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14271,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:hostPort: {} + f:name: {} + f:protocol: {} + f:readinessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/pki/ca-trust/source/service-ca"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + f:dnsPolicy: {} + f:enableServiceLinks: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:serviceAccount: {} + f:serviceAccountName: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"agent-as-daemonset-service-ca"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"agent-as-daemonset-trusted-ca"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + f:updateStrategy: + f:rollingUpdate: + .: {} + f:maxSurge: {} + f:maxUnavailable: {} + f:type: {} + manager: jaeger-operator + operation: Update + time: "2024-01-22T07:16:09Z" + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:status: + f:currentNumberScheduled: {} + f:desiredNumberScheduled: {} + f:numberAvailable: {} + f:numberReady: {} + f:observedGeneration: {} + f:updatedNumberScheduled: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-01-22T07:16:13Z" name: agent-as-daemonset-agent-daemonset namespace: kuttl-test-loved-goldfish + ownerReferences: + - apiVersion: jaegertracing.io/v1 + controller: true + kind: Jaeger + name: agent-as-daemonset + uid: 6fe15c79-36d0-477d-861d-ee90ef391c16 +spec: + revisionHistoryLimit: 10 + selector: + matchLabels: + app: jaeger + app.kubernetes.io/component: agent + app.kubernetes.io/instance: agent-as-daemonset + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: agent-as-daemonset-agent + app.kubernetes.io/part-of: jaeger + template: + metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "14271" + prometheus.io/scrape: "true" + sidecar.istio.io/inject: "false" + creationTimestamp: null + labels: + app: jaeger + app.kubernetes.io/component: agent + app.kubernetes.io/instance: agent-as-daemonset + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: agent-as-daemonset-agent + app.kubernetes.io/part-of: jaeger + spec: + containers: + - args: + - --log-level=debug + - --reporter.grpc.host-port=dns:///agent-as-daemonset-collector-headless.kuttl-test-loved-goldfish:14250 + - --reporter.grpc.tls.ca=/etc/pki/ca-trust/source/service-ca/service-ca.crt + - --reporter.grpc.tls.enabled=true + - --reporter.grpc.tls.server-name=agent-as-daemonset-collector-headless.kuttl-test-loved-goldfish.svc.cluster.local + image: registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79 + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 14271 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-agent-daemonset + ports: + - containerPort: 5775 + hostPort: 5775 + name: zk-compact-trft + protocol: UDP + - containerPort: 5778 + hostPort: 5778 + name: config-rest + protocol: TCP + - containerPort: 6831 + hostPort: 6831 + name: jg-compact-trft + protocol: UDP + - containerPort: 6832 + hostPort: 6832 + name: jg-binary-trft + protocol: UDP + - containerPort: 14271 + hostPort: 14271 + name: admin-http + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 14271 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/pki/ca-trust/extracted/pem + name: agent-as-daemonset-trusted-ca + readOnly: true + - mountPath: /etc/pki/ca-trust/source/service-ca + name: agent-as-daemonset-service-ca + readOnly: true + dnsPolicy: ClusterFirst + enableServiceLinks: false + restartPolicy: Always + schedulerName: default-scheduler + securityContext: {} + serviceAccount: jaeger-agent-daemonset + serviceAccountName: jaeger-agent-daemonset + terminationGracePeriodSeconds: 30 + volumes: + - configMap: + defaultMode: 420 + items: + - key: ca-bundle.crt + path: tls-ca-bundle.pem + name: agent-as-daemonset-trusted-ca + name: agent-as-daemonset-trusted-ca + - configMap: + defaultMode: 420 + items: + - key: service-ca.crt + path: service-ca.crt + name: agent-as-daemonset-service-ca + name: agent-as-daemonset-service-ca + updateStrategy: + rollingUpdate: + maxSurge: 0 + maxUnavailable: 1 + type: RollingUpdate status: - numberReady: 1 + currentNumberScheduled: 3 + desiredNumberScheduled: 3 + numberAvailable: 3 + numberMisscheduled: 0 + numberReady: 3 + observedGeneration: 1 + updatedNumberScheduled: 3 case.go:366: resource DaemonSet:kuttl-test-loved-goldfish/agent-as-daemonset-agent-daemonset: .status.numberReady: value mismatch, expected: 1 != actual: 3 logger.go:42: 07:26:07 | examples-agent-as-daemonset | examples-agent-as-daemonset events from ns kuttl-test-loved-goldfish: logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:09 +0000 UTC Normal Pod agent-as-daemonset-68c47b7f8d-q4hml Binding Scheduled Successfully assigned kuttl-test-loved-goldfish/agent-as-daemonset-68c47b7f8d-q4hml to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:09 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-68c47b7f8d SuccessfulCreate Created pod: agent-as-daemonset-68c47b7f8d-q4hml replicaset-controller logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:09 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-68c47b7f8d to 1 deployment-controller logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:10 +0000 UTC Warning Pod agent-as-daemonset-68c47b7f8d-q4hml FailedMount MountVolume.SetUp failed for volume "agent-as-daemonset-collector-tls-config-volume" : secret "agent-as-daemonset-collector-headless-tls" not found kubelet logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:10 +0000 UTC Warning Pod agent-as-daemonset-68c47b7f8d-q4hml FailedMount MountVolume.SetUp failed for volume "agent-as-daemonset-ui-oauth-proxy-tls" : secret "agent-as-daemonset-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:10 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-62qjz Binding Scheduled Successfully assigned kuttl-test-loved-goldfish/agent-as-daemonset-agent-daemonset-62qjz to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:10 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-62qjz AddedInterface Add eth0 [10.129.2.45/23] from ovn-kubernetes logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:10 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-62qjz.spec.containers{jaeger-agent-daemonset} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:10 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-62qjz.spec.containers{jaeger-agent-daemonset} Created Created container jaeger-agent-daemonset kubelet logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:10 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-62qjz.spec.containers{jaeger-agent-daemonset} Started Started container jaeger-agent-daemonset kubelet logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:10 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-9vxsz Binding Scheduled Successfully assigned kuttl-test-loved-goldfish/agent-as-daemonset-agent-daemonset-9vxsz to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:10 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-z9479 Binding Scheduled Successfully assigned kuttl-test-loved-goldfish/agent-as-daemonset-agent-daemonset-z9479 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:10 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-z9479 AddedInterface Add eth0 [10.128.2.80/23] from ovn-kubernetes logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:10 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-z9479.spec.containers{jaeger-agent-daemonset} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" kubelet logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:10 +0000 UTC Normal DaemonSet.apps agent-as-daemonset-agent-daemonset SuccessfulCreate Created pod: agent-as-daemonset-agent-daemonset-z9479 daemonset-controller logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:10 +0000 UTC Normal DaemonSet.apps agent-as-daemonset-agent-daemonset SuccessfulCreate Created pod: agent-as-daemonset-agent-daemonset-62qjz daemonset-controller logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:10 +0000 UTC Normal DaemonSet.apps agent-as-daemonset-agent-daemonset SuccessfulCreate Created pod: agent-as-daemonset-agent-daemonset-9vxsz daemonset-controller logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:11 +0000 UTC Normal Pod agent-as-daemonset-68c47b7f8d-q4hml AddedInterface Add eth0 [10.131.0.55/23] from ovn-kubernetes logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:11 +0000 UTC Normal Pod agent-as-daemonset-68c47b7f8d-q4hml.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:11 +0000 UTC Normal Pod agent-as-daemonset-68c47b7f8d-q4hml.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:11 +0000 UTC Normal Pod agent-as-daemonset-68c47b7f8d-q4hml.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:11 +0000 UTC Normal Pod agent-as-daemonset-68c47b7f8d-q4hml.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:11 +0000 UTC Normal Pod agent-as-daemonset-68c47b7f8d-q4hml.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:11 +0000 UTC Normal Pod agent-as-daemonset-68c47b7f8d-q4hml.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:11 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-9vxsz AddedInterface Add eth0 [10.131.0.56/23] from ovn-kubernetes logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:11 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-9vxsz.spec.containers{jaeger-agent-daemonset} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:11 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-9vxsz.spec.containers{jaeger-agent-daemonset} Created Created container jaeger-agent-daemonset kubelet logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:11 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-9vxsz.spec.containers{jaeger-agent-daemonset} Started Started container jaeger-agent-daemonset kubelet logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:12 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-z9479.spec.containers{jaeger-agent-daemonset} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" in 1.831s (1.831s including waiting) kubelet logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:12 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-z9479.spec.containers{jaeger-agent-daemonset} Created Created container jaeger-agent-daemonset kubelet logger.go:42: 07:26:07 | examples-agent-as-daemonset | 2024-01-22 07:16:12 +0000 UTC Normal Pod agent-as-daemonset-agent-daemonset-z9479.spec.containers{jaeger-agent-daemonset} Started Started container jaeger-agent-daemonset kubelet logger.go:42: 07:26:07 | examples-agent-as-daemonset | Deleting namespace: kuttl-test-loved-goldfish === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (1498.03s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.85s) --- PASS: kuttl/harness/examples-simple-prod (68.30s) --- PASS: kuttl/harness/examples-with-sampling (312.67s) --- PASS: kuttl/harness/examples-with-cassandra (57.34s) --- PASS: kuttl/harness/examples-with-badger-and-volume (40.04s) --- PASS: kuttl/harness/examples-with-badger (40.37s) --- PASS: kuttl/harness/examples-simplest (34.42s) --- PASS: kuttl/harness/examples-simple-prod-with-volumes (75.49s) --- PASS: kuttl/harness/examples-business-application-injected-sidecar (41.07s) --- PASS: kuttl/harness/examples-service-types (58.72s) --- PASS: kuttl/harness/examples-openshift-with-htpasswd (24.36s) --- PASS: kuttl/harness/examples-collector-with-priority-class (38.66s) --- PASS: kuttl/harness/examples-agent-with-priority-class (53.51s) --- PASS: kuttl/harness/examples-all-in-one-with-options (33.69s) --- FAIL: kuttl/harness/examples-agent-as-daemonset (613.47s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name examples --report --output /logs/artifacts/examples.xml ./artifacts/kuttl-report.xml time="2024-01-22T07:26:14Z" level=debug msg="Setting a new name for the test suites" time="2024-01-22T07:26:14Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-01-22T07:26:14Z" level=debug msg="normalizing test case names" time="2024-01-22T07:26:14Z" level=debug msg="examples/artifacts -> examples_artifacts" time="2024-01-22T07:26:14Z" level=debug msg="examples/examples-simple-prod -> examples_examples_simple_prod" time="2024-01-22T07:26:14Z" level=debug msg="examples/examples-with-sampling -> examples_examples_with_sampling" time="2024-01-22T07:26:14Z" level=debug msg="examples/examples-with-cassandra -> examples_examples_with_cassandra" time="2024-01-22T07:26:14Z" level=debug msg="examples/examples-with-badger-and-volume -> examples_examples_with_badger_and_volume" time="2024-01-22T07:26:14Z" level=debug msg="examples/examples-with-badger -> examples_examples_with_badger" time="2024-01-22T07:26:14Z" level=debug msg="examples/examples-simplest -> examples_examples_simplest" time="2024-01-22T07:26:14Z" level=debug msg="examples/examples-simple-prod-with-volumes -> examples_examples_simple_prod_with_volumes" time="2024-01-22T07:26:14Z" level=debug msg="examples/examples-business-application-injected-sidecar -> examples_examples_business_application_injected_sidecar" time="2024-01-22T07:26:14Z" level=debug msg="examples/examples-service-types -> examples_examples_service_types" time="2024-01-22T07:26:14Z" level=debug msg="examples/examples-openshift-with-htpasswd -> examples_examples_openshift_with_htpasswd" time="2024-01-22T07:26:14Z" level=debug msg="examples/examples-collector-with-priority-class -> examples_examples_collector_with_priority_class" time="2024-01-22T07:26:14Z" level=debug msg="examples/examples-agent-with-priority-class -> examples_examples_agent_with_priority_class" time="2024-01-22T07:26:14Z" level=debug msg="examples/examples-all-in-one-with-options -> examples_examples_all_in_one_with_options" time="2024-01-22T07:26:14Z" level=debug msg="examples/examples-agent-as-daemonset -> examples_examples_agent_as_daemonset" +---------------------------------------------------------+--------+ | NAME | RESULT | +---------------------------------------------------------+--------+ | examples_artifacts | passed | | examples_examples_simple_prod | passed | | examples_examples_with_sampling | passed | | examples_examples_with_cassandra | passed | | examples_examples_with_badger_and_volume | passed | | examples_examples_with_badger | passed | | examples_examples_simplest | passed | | examples_examples_simple_prod_with_volumes | passed | | examples_examples_business_application_injected_sidecar | passed | | examples_examples_service_types | passed | | examples_examples_openshift_with_htpasswd | passed | | examples_examples_collector_with_priority_class | passed | | examples_examples_agent_with_priority_class | passed | | examples_examples_all_in_one_with_options | passed | | examples_examples_agent_as_daemonset | failed | +---------------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh generate false true + '[' 3 -ne 3 ']' + test_suite_name=generate + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/generate.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-generate make[2]: Entering directory '/tmp/jaeger-tests' test -s /tmp/jaeger-tests/bin/operator-sdk || curl -sLo /tmp/jaeger-tests/bin/operator-sdk https://github.com/operator-framework/operator-sdk/releases/download/v1.32.0/operator-sdk_`go env GOOS`_`go env GOARCH` ./hack/install/install-golangci-lint.sh Installing golangci-lint golangci-lint 1.55.2 is installed already ./hack/install/install-goimports.sh Installing goimports Try 0... go install golang.org/x/tools/cmd/goimports@v0.1.12 >>>> Formatting code... ./.ci/format.sh >>>> Building... ./hack/install/install-dependencies.sh Installing go dependencies Try 0... go mod download GOOS= GOARCH= CGO_ENABLED=0 GO111MODULE=on go build -ldflags "-X "github.com/jaegertracing/jaeger-operator/pkg/version".version="1.53.0" -X "github.com/jaegertracing/jaeger-operator/pkg/version".buildDate=2024-01-22T07:26:18Z -X "github.com/jaegertracing/jaeger-operator/pkg/version".defaultJaeger="1.53.0"" -o "bin/jaeger-operator" main.go JAEGER_VERSION="1.53.0" ./tests/e2e/generate/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-18-050837 True False 47m Cluster version is 4.15.0-0.nightly-2024-01-18-050837' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-18-050837 True False 47m Cluster version is 4.15.0-0.nightly-2024-01-18-050837' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/generate/render.sh ++ export SUITE_DIR=./tests/e2e/generate ++ SUITE_DIR=./tests/e2e/generate ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/generate ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test generate 'This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 2 -ne 2 ']' + test_name=generate + message='This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/generate/_build + '[' _build '!=' _build ']' + rm -rf generate + warning 'generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed\e[0m' WAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running generate E2E tests' Running generate E2E tests + cd tests/e2e/generate/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2576283944 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:26:35 | artifacts | Creating namespace: kuttl-test-classic-firefly logger.go:42: 07:26:35 | artifacts | artifacts events from ns kuttl-test-classic-firefly: logger.go:42: 07:26:35 | artifacts | Deleting namespace: kuttl-test-classic-firefly === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (5.91s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.87s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name generate --report --output /logs/artifacts/generate.xml ./artifacts/kuttl-report.xml time="2024-01-22T07:26:41Z" level=debug msg="Setting a new name for the test suites" time="2024-01-22T07:26:41Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-01-22T07:26:41Z" level=debug msg="normalizing test case names" time="2024-01-22T07:26:41Z" level=debug msg="generate/artifacts -> generate_artifacts" +--------------------+--------+ | NAME | RESULT | +--------------------+--------+ | generate_artifacts | passed | +--------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh miscellaneous false true + '[' 3 -ne 3 ']' + test_suite_name=miscellaneous + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/miscellaneous.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-miscellaneous make[2]: Entering directory '/tmp/jaeger-tests' SKIP_ES_EXTERNAL=true ./tests/e2e/miscellaneous/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-18-050837 True False 47m Cluster version is 4.15.0-0.nightly-2024-01-18-050837' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-18-050837 True False 47m Cluster version is 4.15.0-0.nightly-2024-01-18-050837' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/miscellaneous/render.sh ++ export SUITE_DIR=./tests/e2e/miscellaneous ++ SUITE_DIR=./tests/e2e/miscellaneous ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/miscellaneous ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test cassandra-spark 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=cassandra-spark + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf cassandra-spark + warning 'cassandra-spark: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: cassandra-spark: Test not supported in OpenShift\e[0m' WAR: cassandra-spark: Test not supported in OpenShift + start_test collector-autoscale + '[' 1 -ne 1 ']' + test_name=collector-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-autoscale\e[0m' Rendering files for test collector-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p collector-autoscale + cd collector-autoscale + jaeger_name=simple-prod + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + ELASTICSEARCH_NODECOUNT=1 + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.resources.requests.memory="200m"' 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.autoscale=true 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.minReplicas=1 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.maxReplicas=2 01-install.yaml + version_lt 1.27 1.23 ++ echo 1.27 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.27 '!=' 1.27 + rm ./03-assert.yaml + generate_otlp_e2e_tests http + test_protocol=http + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-http\e[0m' Rendering files for test collector-otlp-allinone-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-autoscale + '[' collector-autoscale '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-http + cd collector-otlp-allinone-http + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger http true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-http\e[0m' Rendering files for test collector-otlp-production-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-http + '[' collector-otlp-allinone-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-http + cd collector-otlp-production-http + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger http true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + generate_otlp_e2e_tests grpc + test_protocol=grpc + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-grpc\e[0m' Rendering files for test collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-http + '[' collector-otlp-production-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-grpc + cd collector-otlp-allinone-grpc + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-grpc\e[0m' Rendering files for test collector-otlp-production-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-grpc + '[' collector-otlp-allinone-grpc '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-grpc + cd collector-otlp-production-grpc + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + '[' true = true ']' + skip_test istio 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=istio + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-grpc + '[' collector-otlp-production-grpc '!=' _build ']' + cd .. + rm -rf istio + warning 'istio: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: istio: Test not supported in OpenShift\e[0m' WAR: istio: Test not supported in OpenShift + '[' true = true ']' + skip_test outside-cluster 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=outside-cluster + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf outside-cluster + warning 'outside-cluster: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: outside-cluster: Test not supported in OpenShift\e[0m' WAR: outside-cluster: Test not supported in OpenShift + start_test set-custom-img + '[' 1 -ne 1 ']' + test_name=set-custom-img + echo =========================================================================== =========================================================================== + info 'Rendering files for test set-custom-img' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test set-custom-img\e[0m' Rendering files for test set-custom-img + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p set-custom-img + cd set-custom-img + jaeger_name=my-jaeger + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.image="test"' ./02-install.yaml + '[' true = true ']' + skip_test non-cluster-wide 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=non-cluster-wide + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/set-custom-img + '[' set-custom-img '!=' _build ']' + cd .. + rm -rf non-cluster-wide + warning 'non-cluster-wide: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: non-cluster-wide: Test not supported in OpenShift\e[0m' WAR: non-cluster-wide: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running miscellaneous E2E tests' Running miscellaneous E2E tests + cd tests/e2e/miscellaneous/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2576283944 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/collector-autoscale === PAUSE kuttl/harness/collector-autoscale === RUN kuttl/harness/collector-otlp-allinone-grpc === PAUSE kuttl/harness/collector-otlp-allinone-grpc === RUN kuttl/harness/collector-otlp-allinone-http === PAUSE kuttl/harness/collector-otlp-allinone-http === RUN kuttl/harness/collector-otlp-production-grpc === PAUSE kuttl/harness/collector-otlp-production-grpc === RUN kuttl/harness/collector-otlp-production-http === PAUSE kuttl/harness/collector-otlp-production-http === RUN kuttl/harness/set-custom-img === PAUSE kuttl/harness/set-custom-img === CONT kuttl/harness/artifacts logger.go:42: 07:26:52 | artifacts | Creating namespace: kuttl-test-summary-raptor logger.go:42: 07:26:52 | artifacts | artifacts events from ns kuttl-test-summary-raptor: logger.go:42: 07:26:52 | artifacts | Deleting namespace: kuttl-test-summary-raptor === CONT kuttl/harness/collector-otlp-production-grpc logger.go:42: 07:26:59 | collector-otlp-production-grpc | Creating namespace: kuttl-test-magical-mako logger.go:42: 07:26:59 | collector-otlp-production-grpc/1-install | starting test step 1-install logger.go:42: 07:26:59 | collector-otlp-production-grpc/1-install | Jaeger:kuttl-test-magical-mako/my-jaeger created logger.go:42: 07:27:36 | collector-otlp-production-grpc/1-install | test step completed 1-install logger.go:42: 07:27:36 | collector-otlp-production-grpc/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:27:36 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:27:38 | collector-otlp-production-grpc/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:27:44 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:27:45 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:27:45 | collector-otlp-production-grpc/2-smoke-test | job.batch/report-span created logger.go:42: 07:27:45 | collector-otlp-production-grpc/2-smoke-test | job.batch/check-span created logger.go:42: 07:28:04 | collector-otlp-production-grpc/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:28:04 | collector-otlp-production-grpc | collector-otlp-production-grpc events from ns kuttl-test-magical-mako: logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:04 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestmagicalmakomyjaeger-1-846df6fcf7 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestmagicalmakomyjaeger-1-846df6fcfhlx7f replicaset-controller logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmagicalmakomyjaeger-1-846df6fcfhlx7f Binding Scheduled Successfully assigned kuttl-test-magical-mako/elasticsearch-cdm-kuttltestmagicalmakomyjaeger-1-846df6fcfhlx7f to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:04 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestmagicalmakomyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestmagicalmakomyjaeger-1-846df6fcf7 to 1 deployment-controller logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:05 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmagicalmakomyjaeger-1-846df6fcfhlx7f AddedInterface Add eth0 [10.131.0.57/23] from ovn-kubernetes logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:05 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmagicalmakomyjaeger-1-846df6fcfhlx7f.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:3ad088cbfcfb006d87762e96a70ff989e75d49872954347f7efe34d9529c1b7a" already present on machine kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:05 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmagicalmakomyjaeger-1-846df6fcfhlx7f.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:05 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmagicalmakomyjaeger-1-846df6fcfhlx7f.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:05 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmagicalmakomyjaeger-1-846df6fcfhlx7f.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:911ee3f26ae71eab71bf2a1455d1c2ea871cddc78a51189e3ad2f18f63bd3d91" already present on machine kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:05 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmagicalmakomyjaeger-1-846df6fcfhlx7f.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:05 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmagicalmakomyjaeger-1-846df6fcfhlx7f.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:20 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmagicalmakomyjaeger-1-846df6fcfhlx7f.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:31 +0000 UTC Normal Pod my-jaeger-collector-5499fcb55d-nln5t Binding Scheduled Successfully assigned kuttl-test-magical-mako/my-jaeger-collector-5499fcb55d-nln5t to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:31 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5499fcb55d SuccessfulCreate Created pod: my-jaeger-collector-5499fcb55d-nln5t replicaset-controller logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:31 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-5499fcb55d to 1 deployment-controller logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:31 +0000 UTC Normal Pod my-jaeger-query-69dc9fd89c-kc5xf Binding Scheduled Successfully assigned kuttl-test-magical-mako/my-jaeger-query-69dc9fd89c-kc5xf to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:31 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-69dc9fd89c SuccessfulCreate Created pod: my-jaeger-query-69dc9fd89c-kc5xf replicaset-controller logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:31 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-69dc9fd89c to 1 deployment-controller logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:32 +0000 UTC Warning Pod my-jaeger-collector-5499fcb55d-nln5t FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:32 +0000 UTC Normal Pod my-jaeger-collector-5499fcb55d-nln5t AddedInterface Add eth0 [10.128.2.81/23] from ovn-kubernetes logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:33 +0000 UTC Normal Pod my-jaeger-collector-5499fcb55d-nln5t.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:33 +0000 UTC Normal Pod my-jaeger-collector-5499fcb55d-nln5t.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:33 +0000 UTC Normal Pod my-jaeger-collector-5499fcb55d-nln5t.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:33 +0000 UTC Warning Pod my-jaeger-query-69dc9fd89c-kc5xf FailedMount MountVolume.SetUp failed for volume "certs" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:34 +0000 UTC Normal Pod my-jaeger-query-69dc9fd89c-kc5xf AddedInterface Add eth0 [10.129.2.46/23] from ovn-kubernetes logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:34 +0000 UTC Normal Pod my-jaeger-query-69dc9fd89c-kc5xf.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:34 +0000 UTC Normal Pod my-jaeger-query-69dc9fd89c-kc5xf.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:34 +0000 UTC Normal Pod my-jaeger-query-69dc9fd89c-kc5xf.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:34 +0000 UTC Normal Pod my-jaeger-query-69dc9fd89c-kc5xf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:34 +0000 UTC Normal Pod my-jaeger-query-69dc9fd89c-kc5xf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:34 +0000 UTC Normal Pod my-jaeger-query-69dc9fd89c-kc5xf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:34 +0000 UTC Normal Pod my-jaeger-query-69dc9fd89c-kc5xf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:34 +0000 UTC Normal Pod my-jaeger-query-69dc9fd89c-kc5xf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:34 +0000 UTC Normal Pod my-jaeger-query-69dc9fd89c-kc5xf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:40 +0000 UTC Normal Pod my-jaeger-query-69dc9fd89c-kc5xf.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:40 +0000 UTC Normal Pod my-jaeger-query-69dc9fd89c-kc5xf.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:40 +0000 UTC Normal Pod my-jaeger-query-69dc9fd89c-kc5xf.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:40 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-69dc9fd89c SuccessfulDelete Deleted pod: my-jaeger-query-69dc9fd89c-kc5xf replicaset-controller logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:40 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-69dc9fd89c to 0 from 1 deployment-controller logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:41 +0000 UTC Normal Pod my-jaeger-query-56bf6759b-nb8ps Binding Scheduled Successfully assigned kuttl-test-magical-mako/my-jaeger-query-56bf6759b-nb8ps to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:41 +0000 UTC Normal Pod my-jaeger-query-56bf6759b-nb8ps AddedInterface Add eth0 [10.129.2.47/23] from ovn-kubernetes logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:41 +0000 UTC Normal Pod my-jaeger-query-56bf6759b-nb8ps.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:41 +0000 UTC Normal Pod my-jaeger-query-56bf6759b-nb8ps.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:41 +0000 UTC Normal Pod my-jaeger-query-56bf6759b-nb8ps.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:41 +0000 UTC Normal Pod my-jaeger-query-56bf6759b-nb8ps.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:41 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-56bf6759b SuccessfulCreate Created pod: my-jaeger-query-56bf6759b-nb8ps replicaset-controller logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:41 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-56bf6759b to 1 deployment-controller logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:42 +0000 UTC Normal Pod my-jaeger-query-56bf6759b-nb8ps.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:42 +0000 UTC Normal Pod my-jaeger-query-56bf6759b-nb8ps.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:42 +0000 UTC Normal Pod my-jaeger-query-56bf6759b-nb8ps.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:42 +0000 UTC Normal Pod my-jaeger-query-56bf6759b-nb8ps.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:42 +0000 UTC Normal Pod my-jaeger-query-56bf6759b-nb8ps.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:45 +0000 UTC Normal Pod check-span-sl926 Binding Scheduled Successfully assigned kuttl-test-magical-mako/check-span-sl926 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:45 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-sl926 job-controller logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:45 +0000 UTC Normal Pod report-span-x5jch Binding Scheduled Successfully assigned kuttl-test-magical-mako/report-span-x5jch to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:45 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-x5jch job-controller logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:46 +0000 UTC Normal Pod check-span-sl926 AddedInterface Add eth0 [10.128.2.83/23] from ovn-kubernetes logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:46 +0000 UTC Normal Pod check-span-sl926.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:46 +0000 UTC Normal Pod check-span-sl926.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:46 +0000 UTC Normal Pod check-span-sl926.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:46 +0000 UTC Normal Pod report-span-x5jch AddedInterface Add eth0 [10.128.2.82/23] from ovn-kubernetes logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:46 +0000 UTC Normal Pod report-span-x5jch.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:46 +0000 UTC Normal Pod report-span-x5jch.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:46 +0000 UTC Normal Pod report-span-x5jch.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:47 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:47 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:27:47 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:28:04 | collector-otlp-production-grpc | 2024-01-22 07:28:04 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:28:04 | collector-otlp-production-grpc | Deleting namespace: kuttl-test-magical-mako === CONT kuttl/harness/set-custom-img logger.go:42: 07:28:17 | set-custom-img | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:28:17 | set-custom-img | Ignoring check-collector-img.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:28:17 | set-custom-img | Creating namespace: kuttl-test-useful-lacewing logger.go:42: 07:28:17 | set-custom-img/1-install | starting test step 1-install logger.go:42: 07:28:17 | set-custom-img/1-install | Jaeger:kuttl-test-useful-lacewing/my-jaeger created logger.go:42: 07:28:54 | set-custom-img/1-install | test step completed 1-install logger.go:42: 07:28:54 | set-custom-img/2-install | starting test step 2-install logger.go:42: 07:28:54 | set-custom-img/2-install | Jaeger:kuttl-test-useful-lacewing/my-jaeger updated logger.go:42: 07:28:54 | set-custom-img/2-install | test step completed 2-install logger.go:42: 07:28:54 | set-custom-img/3-check-image | starting test step 3-check-image logger.go:42: 07:28:54 | set-custom-img/3-check-image | running command: [sh -c ./check-collector-img.sh] logger.go:42: 07:28:54 | set-custom-img/3-check-image | Collector image missmatch. Expected: test. Has: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9 logger.go:42: 07:28:59 | set-custom-img/3-check-image | Collector image asserted properly! logger.go:42: 07:28:59 | set-custom-img/3-check-image | test step completed 3-check-image logger.go:42: 07:28:59 | set-custom-img | set-custom-img events from ns kuttl-test-useful-lacewing: logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:22 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestusefullacewingmyjaeger-1-dcb649fc9 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestusefullacewingmyjaeger-1-dcb649rz55d replicaset-controller logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestusefullacewingmyjaeger-1-dcb649rz55d Binding Scheduled Successfully assigned kuttl-test-useful-lacewing/elasticsearch-cdm-kuttltestusefullacewingmyjaeger-1-dcb649rz55d to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:22 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestusefullacewingmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestusefullacewingmyjaeger-1-dcb649fc9 to 1 deployment-controller logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:24 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestusefullacewingmyjaeger-1-dcb649rz55d FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestusefullacewingmyjaeger-1-dcb649rz55d AddedInterface Add eth0 [10.131.0.58/23] from ovn-kubernetes logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestusefullacewingmyjaeger-1-dcb649rz55d.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:3ad088cbfcfb006d87762e96a70ff989e75d49872954347f7efe34d9529c1b7a" already present on machine kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestusefullacewingmyjaeger-1-dcb649rz55d.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestusefullacewingmyjaeger-1-dcb649rz55d.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestusefullacewingmyjaeger-1-dcb649rz55d.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:911ee3f26ae71eab71bf2a1455d1c2ea871cddc78a51189e3ad2f18f63bd3d91" already present on machine kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestusefullacewingmyjaeger-1-dcb649rz55d.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestusefullacewingmyjaeger-1-dcb649rz55d.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:39 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestusefullacewingmyjaeger-1-dcb649rz55d.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:50 +0000 UTC Normal Pod my-jaeger-collector-7dc87d5db-bgkl8 Binding Scheduled Successfully assigned kuttl-test-useful-lacewing/my-jaeger-collector-7dc87d5db-bgkl8 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:50 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7dc87d5db SuccessfulCreate Created pod: my-jaeger-collector-7dc87d5db-bgkl8 replicaset-controller logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:50 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7dc87d5db to 1 deployment-controller logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:50 +0000 UTC Normal Pod my-jaeger-query-cdb677f6f-xpvgv Binding Scheduled Successfully assigned kuttl-test-useful-lacewing/my-jaeger-query-cdb677f6f-xpvgv to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:50 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-cdb677f6f SuccessfulCreate Created pod: my-jaeger-query-cdb677f6f-xpvgv replicaset-controller logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:50 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-cdb677f6f to 1 deployment-controller logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:51 +0000 UTC Normal Pod my-jaeger-collector-7dc87d5db-bgkl8 AddedInterface Add eth0 [10.128.2.84/23] from ovn-kubernetes logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:51 +0000 UTC Normal Pod my-jaeger-collector-7dc87d5db-bgkl8.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:51 +0000 UTC Normal Pod my-jaeger-collector-7dc87d5db-bgkl8.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:51 +0000 UTC Normal Pod my-jaeger-collector-7dc87d5db-bgkl8.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:52 +0000 UTC Warning Pod my-jaeger-query-cdb677f6f-xpvgv FailedMount MountVolume.SetUp failed for volume "my-jaeger-service-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:52 +0000 UTC Warning Pod my-jaeger-query-cdb677f6f-xpvgv FailedMount MountVolume.SetUp failed for volume "certs" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:52 +0000 UTC Warning Pod my-jaeger-query-cdb677f6f-xpvgv FailedMount MountVolume.SetUp failed for volume "kube-api-access-6bqsb" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:53 +0000 UTC Normal Pod my-jaeger-query-cdb677f6f-xpvgv AddedInterface Add eth0 [10.129.2.48/23] from ovn-kubernetes logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:53 +0000 UTC Normal Pod my-jaeger-query-cdb677f6f-xpvgv.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:53 +0000 UTC Normal Pod my-jaeger-query-cdb677f6f-xpvgv.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:53 +0000 UTC Normal Pod my-jaeger-query-cdb677f6f-xpvgv.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:53 +0000 UTC Normal Pod my-jaeger-query-cdb677f6f-xpvgv.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:53 +0000 UTC Normal Pod my-jaeger-query-cdb677f6f-xpvgv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:53 +0000 UTC Normal Pod my-jaeger-query-cdb677f6f-xpvgv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:53 +0000 UTC Normal Pod my-jaeger-query-cdb677f6f-xpvgv.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:53 +0000 UTC Normal Pod my-jaeger-query-cdb677f6f-xpvgv.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:53 +0000 UTC Normal Pod my-jaeger-query-cdb677f6f-xpvgv.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:56 +0000 UTC Normal Pod my-jaeger-collector-7dc87d5db-bgkl8.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:56 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7dc87d5db SuccessfulDelete Deleted pod: my-jaeger-collector-7dc87d5db-bgkl8 replicaset-controller logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:56 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-7dc87d5db to 0 from 1 deployment-controller logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:57 +0000 UTC Normal Pod my-jaeger-collector-78dcf9bd6d-kbpsh Binding Scheduled Successfully assigned kuttl-test-useful-lacewing/my-jaeger-collector-78dcf9bd6d-kbpsh to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:57 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-78dcf9bd6d SuccessfulCreate Created pod: my-jaeger-collector-78dcf9bd6d-kbpsh replicaset-controller logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:57 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-78dcf9bd6d to 1 deployment-controller logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:58 +0000 UTC Normal Pod my-jaeger-collector-78dcf9bd6d-kbpsh AddedInterface Add eth0 [10.128.2.85/23] from ovn-kubernetes logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:58 +0000 UTC Normal Pod my-jaeger-collector-78dcf9bd6d-kbpsh.spec.containers{jaeger-collector} Pulling Pulling image "test" kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:58 +0000 UTC Warning Pod my-jaeger-collector-78dcf9bd6d-kbpsh.spec.containers{jaeger-collector} Failed Failed to pull image "test": reading manifest latest in docker.io/library/test: requested access to the resource is denied kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:58 +0000 UTC Warning Pod my-jaeger-collector-78dcf9bd6d-kbpsh.spec.containers{jaeger-collector} Failed Error: ErrImagePull kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:58 +0000 UTC Normal Pod my-jaeger-collector-78dcf9bd6d-kbpsh.spec.containers{jaeger-collector} BackOff Back-off pulling image "test" kubelet logger.go:42: 07:28:59 | set-custom-img | 2024-01-22 07:28:58 +0000 UTC Warning Pod my-jaeger-collector-78dcf9bd6d-kbpsh.spec.containers{jaeger-collector} Failed Error: ImagePullBackOff kubelet logger.go:42: 07:28:59 | set-custom-img | Deleting namespace: kuttl-test-useful-lacewing === CONT kuttl/harness/collector-otlp-production-http logger.go:42: 07:29:06 | collector-otlp-production-http | Creating namespace: kuttl-test-flowing-lamb logger.go:42: 07:29:06 | collector-otlp-production-http/1-install | starting test step 1-install logger.go:42: 07:29:06 | collector-otlp-production-http/1-install | Jaeger:kuttl-test-flowing-lamb/my-jaeger created logger.go:42: 07:29:43 | collector-otlp-production-http/1-install | test step completed 1-install logger.go:42: 07:29:43 | collector-otlp-production-http/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:29:43 | collector-otlp-production-http/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:29:44 | collector-otlp-production-http/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:29:53 | collector-otlp-production-http/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:29:53 | collector-otlp-production-http/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:29:54 | collector-otlp-production-http/2-smoke-test | job.batch/report-span created logger.go:42: 07:29:54 | collector-otlp-production-http/2-smoke-test | job.batch/check-span created logger.go:42: 07:30:06 | collector-otlp-production-http/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:30:06 | collector-otlp-production-http | collector-otlp-production-http events from ns kuttl-test-flowing-lamb: logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:13 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestflowinglambmyjaeger-1-668896f9c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestflowinglambmyjaeger-1-668896f9cdt5bq replicaset-controller logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:13 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestflowinglambmyjaeger-1-668896f9cdt5bq Binding Scheduled Successfully assigned kuttl-test-flowing-lamb/elasticsearch-cdm-kuttltestflowinglambmyjaeger-1-668896f9cdt5bq to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:13 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestflowinglambmyjaeger-1-668896f9cdt5bq AddedInterface Add eth0 [10.131.0.59/23] from ovn-kubernetes logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:13 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestflowinglambmyjaeger-1-668896f9cdt5bq.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:3ad088cbfcfb006d87762e96a70ff989e75d49872954347f7efe34d9529c1b7a" already present on machine kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:13 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestflowinglambmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestflowinglambmyjaeger-1-668896f9c to 1 deployment-controller logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestflowinglambmyjaeger-1-668896f9cdt5bq.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestflowinglambmyjaeger-1-668896f9cdt5bq.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestflowinglambmyjaeger-1-668896f9cdt5bq.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:911ee3f26ae71eab71bf2a1455d1c2ea871cddc78a51189e3ad2f18f63bd3d91" already present on machine kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestflowinglambmyjaeger-1-668896f9cdt5bq.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestflowinglambmyjaeger-1-668896f9cdt5bq.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:28 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestflowinglambmyjaeger-1-668896f9cdt5bq.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:40 +0000 UTC Normal Pod my-jaeger-collector-f69f586d4-vqm7d Binding Scheduled Successfully assigned kuttl-test-flowing-lamb/my-jaeger-collector-f69f586d4-vqm7d to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:40 +0000 UTC Warning Pod my-jaeger-collector-f69f586d4-vqm7d FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:40 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-f69f586d4 SuccessfulCreate Created pod: my-jaeger-collector-f69f586d4-vqm7d replicaset-controller logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:40 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-f69f586d4 to 1 deployment-controller logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:40 +0000 UTC Normal Pod my-jaeger-query-7946b9cdb8-ghclt Binding Scheduled Successfully assigned kuttl-test-flowing-lamb/my-jaeger-query-7946b9cdb8-ghclt to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:40 +0000 UTC Normal Pod my-jaeger-query-7946b9cdb8-ghclt AddedInterface Add eth0 [10.129.2.49/23] from ovn-kubernetes logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:40 +0000 UTC Normal Pod my-jaeger-query-7946b9cdb8-ghclt.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:40 +0000 UTC Normal Pod my-jaeger-query-7946b9cdb8-ghclt.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:40 +0000 UTC Normal Pod my-jaeger-query-7946b9cdb8-ghclt.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:40 +0000 UTC Normal Pod my-jaeger-query-7946b9cdb8-ghclt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:40 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7946b9cdb8 SuccessfulCreate Created pod: my-jaeger-query-7946b9cdb8-ghclt replicaset-controller logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:40 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7946b9cdb8 to 1 deployment-controller logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:41 +0000 UTC Normal Pod my-jaeger-collector-f69f586d4-vqm7d AddedInterface Add eth0 [10.128.2.86/23] from ovn-kubernetes logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:41 +0000 UTC Normal Pod my-jaeger-collector-f69f586d4-vqm7d.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:41 +0000 UTC Normal Pod my-jaeger-collector-f69f586d4-vqm7d.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:41 +0000 UTC Normal Pod my-jaeger-collector-f69f586d4-vqm7d.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:41 +0000 UTC Normal Pod my-jaeger-query-7946b9cdb8-ghclt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:41 +0000 UTC Normal Pod my-jaeger-query-7946b9cdb8-ghclt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:41 +0000 UTC Normal Pod my-jaeger-query-7946b9cdb8-ghclt.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:41 +0000 UTC Normal Pod my-jaeger-query-7946b9cdb8-ghclt.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:41 +0000 UTC Normal Pod my-jaeger-query-7946b9cdb8-ghclt.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:49 +0000 UTC Normal Pod my-jaeger-query-7946b9cdb8-ghclt.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:49 +0000 UTC Normal Pod my-jaeger-query-7946b9cdb8-ghclt.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:49 +0000 UTC Normal Pod my-jaeger-query-7946b9cdb8-ghclt.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:49 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7946b9cdb8 SuccessfulDelete Deleted pod: my-jaeger-query-7946b9cdb8-ghclt replicaset-controller logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:49 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-7946b9cdb8 to 0 from 1 deployment-controller logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:50 +0000 UTC Normal Pod my-jaeger-query-547d57bd8f-n9rcn Binding Scheduled Successfully assigned kuttl-test-flowing-lamb/my-jaeger-query-547d57bd8f-n9rcn to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:50 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-547d57bd8f SuccessfulCreate Created pod: my-jaeger-query-547d57bd8f-n9rcn replicaset-controller logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:50 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-547d57bd8f to 1 deployment-controller logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:51 +0000 UTC Normal Pod my-jaeger-query-547d57bd8f-n9rcn AddedInterface Add eth0 [10.129.2.50/23] from ovn-kubernetes logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:51 +0000 UTC Normal Pod my-jaeger-query-547d57bd8f-n9rcn.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:51 +0000 UTC Normal Pod my-jaeger-query-547d57bd8f-n9rcn.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:51 +0000 UTC Normal Pod my-jaeger-query-547d57bd8f-n9rcn.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:51 +0000 UTC Normal Pod my-jaeger-query-547d57bd8f-n9rcn.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:51 +0000 UTC Normal Pod my-jaeger-query-547d57bd8f-n9rcn.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:51 +0000 UTC Normal Pod my-jaeger-query-547d57bd8f-n9rcn.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:51 +0000 UTC Normal Pod my-jaeger-query-547d57bd8f-n9rcn.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:51 +0000 UTC Normal Pod my-jaeger-query-547d57bd8f-n9rcn.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:51 +0000 UTC Normal Pod my-jaeger-query-547d57bd8f-n9rcn.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:54 +0000 UTC Normal Pod check-span-7627g Binding Scheduled Successfully assigned kuttl-test-flowing-lamb/check-span-7627g to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:54 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-7627g job-controller logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:54 +0000 UTC Normal Pod report-span-72dzq Binding Scheduled Successfully assigned kuttl-test-flowing-lamb/report-span-72dzq to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:54 +0000 UTC Normal Pod report-span-72dzq AddedInterface Add eth0 [10.128.2.87/23] from ovn-kubernetes logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:54 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-72dzq job-controller logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:55 +0000 UTC Normal Pod check-span-7627g AddedInterface Add eth0 [10.128.2.88/23] from ovn-kubernetes logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:55 +0000 UTC Normal Pod check-span-7627g.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:55 +0000 UTC Normal Pod check-span-7627g.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:55 +0000 UTC Normal Pod check-span-7627g.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:55 +0000 UTC Normal Pod report-span-72dzq.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:55 +0000 UTC Normal Pod report-span-72dzq.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:55 +0000 UTC Normal Pod report-span-72dzq.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:29:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:30:06 | collector-otlp-production-http | 2024-01-22 07:30:05 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:30:06 | collector-otlp-production-http | Deleting namespace: kuttl-test-flowing-lamb === CONT kuttl/harness/collector-otlp-allinone-grpc logger.go:42: 07:30:14 | collector-otlp-allinone-grpc | Creating namespace: kuttl-test-relaxed-urchin logger.go:42: 07:30:14 | collector-otlp-allinone-grpc/0-install | starting test step 0-install logger.go:42: 07:30:14 | collector-otlp-allinone-grpc/0-install | Jaeger:kuttl-test-relaxed-urchin/my-jaeger created logger.go:42: 07:30:21 | collector-otlp-allinone-grpc/0-install | test step completed 0-install logger.go:42: 07:30:21 | collector-otlp-allinone-grpc/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:30:21 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:30:22 | collector-otlp-allinone-grpc/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:30:29 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:30:29 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:30:29 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/report-span created logger.go:42: 07:30:29 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/check-span created logger.go:42: 07:30:50 | collector-otlp-allinone-grpc/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | collector-otlp-allinone-grpc events from ns kuttl-test-relaxed-urchin: logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:17 +0000 UTC Normal Pod my-jaeger-5b8c79559f-jljpj Binding Scheduled Successfully assigned kuttl-test-relaxed-urchin/my-jaeger-5b8c79559f-jljpj to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:17 +0000 UTC Warning Pod my-jaeger-5b8c79559f-jljpj FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:17 +0000 UTC Normal ReplicaSet.apps my-jaeger-5b8c79559f SuccessfulCreate Created pod: my-jaeger-5b8c79559f-jljpj replicaset-controller logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:17 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5b8c79559f to 1 deployment-controller logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:18 +0000 UTC Normal Pod my-jaeger-5b8c79559f-jljpj AddedInterface Add eth0 [10.131.0.60/23] from ovn-kubernetes logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:18 +0000 UTC Normal Pod my-jaeger-5b8c79559f-jljpj.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:19 +0000 UTC Normal Pod my-jaeger-5b8c79559f-jljpj.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:19 +0000 UTC Normal Pod my-jaeger-5b8c79559f-jljpj.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:19 +0000 UTC Normal Pod my-jaeger-5b8c79559f-jljpj.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:19 +0000 UTC Normal Pod my-jaeger-5b8c79559f-jljpj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:19 +0000 UTC Normal Pod my-jaeger-5b8c79559f-jljpj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:24 +0000 UTC Normal Pod my-jaeger-5b8c79559f-jljpj.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:24 +0000 UTC Normal Pod my-jaeger-5b8c79559f-jljpj.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:24 +0000 UTC Normal ReplicaSet.apps my-jaeger-5b8c79559f SuccessfulDelete Deleted pod: my-jaeger-5b8c79559f-jljpj replicaset-controller logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:24 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-5b8c79559f to 0 from 1 deployment-controller logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:25 +0000 UTC Normal Pod my-jaeger-6d74cf4f7f-lsx9z Binding Scheduled Successfully assigned kuttl-test-relaxed-urchin/my-jaeger-6d74cf4f7f-lsx9z to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:25 +0000 UTC Normal ReplicaSet.apps my-jaeger-6d74cf4f7f SuccessfulCreate Created pod: my-jaeger-6d74cf4f7f-lsx9z replicaset-controller logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:25 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6d74cf4f7f to 1 deployment-controller logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:26 +0000 UTC Normal Pod my-jaeger-6d74cf4f7f-lsx9z AddedInterface Add eth0 [10.131.0.61/23] from ovn-kubernetes logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:26 +0000 UTC Normal Pod my-jaeger-6d74cf4f7f-lsx9z.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:26 +0000 UTC Normal Pod my-jaeger-6d74cf4f7f-lsx9z.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:26 +0000 UTC Normal Pod my-jaeger-6d74cf4f7f-lsx9z.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:26 +0000 UTC Normal Pod my-jaeger-6d74cf4f7f-lsx9z.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:26 +0000 UTC Normal Pod my-jaeger-6d74cf4f7f-lsx9z.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:26 +0000 UTC Normal Pod my-jaeger-6d74cf4f7f-lsx9z.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:29 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-md5d8 job-controller logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:29 +0000 UTC Normal Pod report-span-4tt48 Binding Scheduled Successfully assigned kuttl-test-relaxed-urchin/report-span-4tt48 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:29 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-4tt48 job-controller logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:30 +0000 UTC Normal Pod check-span-md5d8 Binding Scheduled Successfully assigned kuttl-test-relaxed-urchin/check-span-md5d8 to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:30 +0000 UTC Normal Pod check-span-md5d8 AddedInterface Add eth0 [10.129.2.51/23] from ovn-kubernetes logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:30 +0000 UTC Normal Pod check-span-md5d8.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:30 +0000 UTC Normal Pod check-span-md5d8.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:30 +0000 UTC Normal Pod check-span-md5d8.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:30 +0000 UTC Normal Pod report-span-4tt48 AddedInterface Add eth0 [10.128.2.90/23] from ovn-kubernetes logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:30 +0000 UTC Normal Pod report-span-4tt48.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:30 +0000 UTC Normal Pod report-span-4tt48.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:30 +0000 UTC Normal Pod report-span-4tt48.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | 2024-01-22 07:30:49 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:30:50 | collector-otlp-allinone-grpc | Deleting namespace: kuttl-test-relaxed-urchin === CONT kuttl/harness/collector-otlp-allinone-http logger.go:42: 07:31:02 | collector-otlp-allinone-http | Creating namespace: kuttl-test-star-leech logger.go:42: 07:31:02 | collector-otlp-allinone-http/0-install | starting test step 0-install logger.go:42: 07:31:02 | collector-otlp-allinone-http/0-install | Jaeger:kuttl-test-star-leech/my-jaeger created logger.go:42: 07:31:10 | collector-otlp-allinone-http/0-install | test step completed 0-install logger.go:42: 07:31:10 | collector-otlp-allinone-http/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:31:10 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:31:11 | collector-otlp-allinone-http/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:31:17 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882 OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:31:18 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:31:18 | collector-otlp-allinone-http/1-smoke-test | job.batch/report-span created logger.go:42: 07:31:18 | collector-otlp-allinone-http/1-smoke-test | job.batch/check-span created logger.go:42: 07:31:29 | collector-otlp-allinone-http/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:31:29 | collector-otlp-allinone-http | collector-otlp-allinone-http events from ns kuttl-test-star-leech: logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:05 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-65f7877bc8 to 1 deployment-controller logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:06 +0000 UTC Normal Pod my-jaeger-65f7877bc8-4x8w5 Binding Scheduled Successfully assigned kuttl-test-star-leech/my-jaeger-65f7877bc8-4x8w5 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:06 +0000 UTC Normal ReplicaSet.apps my-jaeger-65f7877bc8 SuccessfulCreate Created pod: my-jaeger-65f7877bc8-4x8w5 replicaset-controller logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:07 +0000 UTC Warning Pod my-jaeger-65f7877bc8-4x8w5 FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-oauth-proxy-tls" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:07 +0000 UTC Warning Pod my-jaeger-65f7877bc8-4x8w5 FailedMount MountVolume.SetUp failed for volume "my-jaeger-trusted-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:07 +0000 UTC Warning Pod my-jaeger-65f7877bc8-4x8w5 FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:07 +0000 UTC Warning Pod my-jaeger-65f7877bc8-4x8w5 FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-configuration-volume" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:08 +0000 UTC Normal Pod my-jaeger-65f7877bc8-4x8w5 AddedInterface Add eth0 [10.131.0.62/23] from ovn-kubernetes logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:08 +0000 UTC Normal Pod my-jaeger-65f7877bc8-4x8w5.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:08 +0000 UTC Normal Pod my-jaeger-65f7877bc8-4x8w5.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:08 +0000 UTC Normal Pod my-jaeger-65f7877bc8-4x8w5.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:08 +0000 UTC Normal Pod my-jaeger-65f7877bc8-4x8w5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:08 +0000 UTC Normal Pod my-jaeger-65f7877bc8-4x8w5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:08 +0000 UTC Normal Pod my-jaeger-65f7877bc8-4x8w5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:12 +0000 UTC Normal Pod my-jaeger-65f7877bc8-4x8w5.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:12 +0000 UTC Normal Pod my-jaeger-65f7877bc8-4x8w5.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:12 +0000 UTC Normal ReplicaSet.apps my-jaeger-65f7877bc8 SuccessfulDelete Deleted pod: my-jaeger-65f7877bc8-4x8w5 replicaset-controller logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:12 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-65f7877bc8 to 0 from 1 deployment-controller logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:13 +0000 UTC Normal ReplicaSet.apps my-jaeger-5f8f4f844c SuccessfulCreate Created pod: my-jaeger-5f8f4f844c-8xs4h replicaset-controller logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:13 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5f8f4f844c to 1 deployment-controller logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:14 +0000 UTC Normal Pod my-jaeger-5f8f4f844c-8xs4h Binding Scheduled Successfully assigned kuttl-test-star-leech/my-jaeger-5f8f4f844c-8xs4h to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:14 +0000 UTC Normal Pod my-jaeger-5f8f4f844c-8xs4h AddedInterface Add eth0 [10.131.0.63/23] from ovn-kubernetes logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:14 +0000 UTC Normal Pod my-jaeger-5f8f4f844c-8xs4h.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:14 +0000 UTC Normal Pod my-jaeger-5f8f4f844c-8xs4h.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:14 +0000 UTC Normal Pod my-jaeger-5f8f4f844c-8xs4h.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:14 +0000 UTC Normal Pod my-jaeger-5f8f4f844c-8xs4h.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:14 +0000 UTC Normal Pod my-jaeger-5f8f4f844c-8xs4h.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:14 +0000 UTC Normal Pod my-jaeger-5f8f4f844c-8xs4h.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:18 +0000 UTC Normal Pod check-span-w7b5s Binding Scheduled Successfully assigned kuttl-test-star-leech/check-span-w7b5s to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:18 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-w7b5s job-controller logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:18 +0000 UTC Normal Pod report-span-n8f44 Binding Scheduled Successfully assigned kuttl-test-star-leech/report-span-n8f44 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:18 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-n8f44 job-controller logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:19 +0000 UTC Normal Pod check-span-w7b5s AddedInterface Add eth0 [10.129.2.52/23] from ovn-kubernetes logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:19 +0000 UTC Normal Pod check-span-w7b5s.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:19 +0000 UTC Normal Pod check-span-w7b5s.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:19 +0000 UTC Normal Pod check-span-w7b5s.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:19 +0000 UTC Normal Pod report-span-n8f44 AddedInterface Add eth0 [10.128.2.91/23] from ovn-kubernetes logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:19 +0000 UTC Normal Pod report-span-n8f44.spec.containers{report-span} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:19 +0000 UTC Normal Pod report-span-n8f44.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:19 +0000 UTC Normal Pod report-span-n8f44.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:31:29 | collector-otlp-allinone-http | 2024-01-22 07:31:29 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:31:29 | collector-otlp-allinone-http | Deleting namespace: kuttl-test-star-leech === CONT kuttl/harness/collector-autoscale logger.go:42: 07:31:41 | collector-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:31:41 | collector-autoscale | Creating namespace: kuttl-test-knowing-vervet logger.go:42: 07:31:41 | collector-autoscale/1-install | starting test step 1-install logger.go:42: 07:31:41 | collector-autoscale/1-install | Jaeger:kuttl-test-knowing-vervet/simple-prod created logger.go:42: 07:32:17 | collector-autoscale/1-install | test step completed 1-install logger.go:42: 07:32:17 | collector-autoscale/2- | starting test step 2- logger.go:42: 07:32:17 | collector-autoscale/2- | test step completed 2- logger.go:42: 07:32:17 | collector-autoscale | collector-autoscale events from ns kuttl-test-knowing-vervet: logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:31:47 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestknowingvervetsimpleprod-1-65dcc45575 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestknowingvervetsimpleprod-1-65dccqsbkd replicaset-controller logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:31:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestknowingvervetsimpleprod-1-65dccqsbkd Binding Scheduled Successfully assigned kuttl-test-knowing-vervet/elasticsearch-cdm-kuttltestknowingvervetsimpleprod-1-65dccqsbkd to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:31:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestknowingvervetsimpleprod-1-65dccqsbkd AddedInterface Add eth0 [10.131.0.64/23] from ovn-kubernetes logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:31:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestknowingvervetsimpleprod-1-65dccqsbkd.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:3ad088cbfcfb006d87762e96a70ff989e75d49872954347f7efe34d9529c1b7a" already present on machine kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:31:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestknowingvervetsimpleprod-1-65dccqsbkd.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:31:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestknowingvervetsimpleprod-1-65dccqsbkd.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:31:47 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestknowingvervetsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestknowingvervetsimpleprod-1-65dcc45575 to 1 deployment-controller logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:31:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestknowingvervetsimpleprod-1-65dccqsbkd.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:911ee3f26ae71eab71bf2a1455d1c2ea871cddc78a51189e3ad2f18f63bd3d91" already present on machine kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:31:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestknowingvervetsimpleprod-1-65dccqsbkd.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:31:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestknowingvervetsimpleprod-1-65dccqsbkd.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:31:57 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestknowingvervetsimpleprod-1-65dccqsbkd.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:02 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestknowingvervetsimpleprod-1-65dccqsbkd.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:14 +0000 UTC Normal Pod simple-prod-collector-fbc7f877c-rrm54 Binding Scheduled Successfully assigned kuttl-test-knowing-vervet/simple-prod-collector-fbc7f877c-rrm54 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:14 +0000 UTC Warning Pod simple-prod-collector-fbc7f877c-rrm54 FailedMount MountVolume.SetUp failed for volume "simple-prod-collector-tls-config-volume" : secret "simple-prod-collector-headless-tls" not found kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:14 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-fbc7f877c SuccessfulCreate Created pod: simple-prod-collector-fbc7f877c-rrm54 replicaset-controller logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:14 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-fbc7f877c to 1 deployment-controller logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:14 +0000 UTC Normal Pod simple-prod-query-8b44f894d-qmtl7 Binding Scheduled Successfully assigned kuttl-test-knowing-vervet/simple-prod-query-8b44f894d-qmtl7 to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:14 +0000 UTC Normal Pod simple-prod-query-8b44f894d-qmtl7 AddedInterface Add eth0 [10.129.2.53/23] from ovn-kubernetes logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:14 +0000 UTC Normal Pod simple-prod-query-8b44f894d-qmtl7.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:14 +0000 UTC Normal ReplicaSet.apps simple-prod-query-8b44f894d SuccessfulCreate Created pod: simple-prod-query-8b44f894d-qmtl7 replicaset-controller logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:14 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-8b44f894d to 1 deployment-controller logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:15 +0000 UTC Normal Pod simple-prod-collector-fbc7f877c-rrm54 AddedInterface Add eth0 [10.128.2.92/23] from ovn-kubernetes logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:15 +0000 UTC Normal Pod simple-prod-collector-fbc7f877c-rrm54.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:15 +0000 UTC Normal Pod simple-prod-collector-fbc7f877c-rrm54.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:15 +0000 UTC Normal Pod simple-prod-collector-fbc7f877c-rrm54.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:15 +0000 UTC Normal Pod simple-prod-query-8b44f894d-qmtl7.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:15 +0000 UTC Normal Pod simple-prod-query-8b44f894d-qmtl7.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:15 +0000 UTC Normal Pod simple-prod-query-8b44f894d-qmtl7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:15 +0000 UTC Normal Pod simple-prod-query-8b44f894d-qmtl7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:15 +0000 UTC Normal Pod simple-prod-query-8b44f894d-qmtl7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:15 +0000 UTC Normal Pod simple-prod-query-8b44f894d-qmtl7.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:15 +0000 UTC Normal Pod simple-prod-query-8b44f894d-qmtl7.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:32:17 | collector-autoscale | 2024-01-22 07:32:15 +0000 UTC Normal Pod simple-prod-query-8b44f894d-qmtl7.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:32:17 | collector-autoscale | Deleting namespace: kuttl-test-knowing-vervet === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (330.97s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.47s) --- PASS: kuttl/harness/collector-otlp-production-grpc (77.97s) --- PASS: kuttl/harness/set-custom-img (48.95s) --- PASS: kuttl/harness/collector-otlp-production-http (68.01s) --- PASS: kuttl/harness/collector-otlp-allinone-grpc (48.08s) --- PASS: kuttl/harness/collector-otlp-allinone-http (39.71s) --- PASS: kuttl/harness/collector-autoscale (41.74s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name miscellaneous --report --output /logs/artifacts/miscellaneous.xml ./artifacts/kuttl-report.xml time="2024-01-22T07:32:24Z" level=debug msg="Setting a new name for the test suites" time="2024-01-22T07:32:24Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-01-22T07:32:24Z" level=debug msg="normalizing test case names" time="2024-01-22T07:32:24Z" level=debug msg="miscellaneous/artifacts -> miscellaneous_artifacts" time="2024-01-22T07:32:24Z" level=debug msg="miscellaneous/collector-otlp-production-grpc -> miscellaneous_collector_otlp_production_grpc" time="2024-01-22T07:32:24Z" level=debug msg="miscellaneous/set-custom-img -> miscellaneous_set_custom_img" time="2024-01-22T07:32:24Z" level=debug msg="miscellaneous/collector-otlp-production-http -> miscellaneous_collector_otlp_production_http" time="2024-01-22T07:32:24Z" level=debug msg="miscellaneous/collector-otlp-allinone-grpc -> miscellaneous_collector_otlp_allinone_grpc" time="2024-01-22T07:32:24Z" level=debug msg="miscellaneous/collector-otlp-allinone-http -> miscellaneous_collector_otlp_allinone_http" time="2024-01-22T07:32:24Z" level=debug msg="miscellaneous/collector-autoscale -> miscellaneous_collector_autoscale" +----------------------------------------------+--------+ | NAME | RESULT | +----------------------------------------------+--------+ | miscellaneous_artifacts | passed | | miscellaneous_collector_otlp_production_grpc | passed | | miscellaneous_set_custom_img | passed | | miscellaneous_collector_otlp_production_http | passed | | miscellaneous_collector_otlp_allinone_grpc | passed | | miscellaneous_collector_otlp_allinone_http | passed | | miscellaneous_collector_autoscale | passed | +----------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh sidecar false true + '[' 3 -ne 3 ']' + test_suite_name=sidecar + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/sidecar.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-sidecar make[2]: Entering directory '/tmp/jaeger-tests' ./tests/e2e/sidecar/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-18-050837 True False 53m Cluster version is 4.15.0-0.nightly-2024-01-18-050837' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-18-050837 True False 53m Cluster version is 4.15.0-0.nightly-2024-01-18-050837' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/sidecar/render.sh ++ export SUITE_DIR=./tests/e2e/sidecar ++ SUITE_DIR=./tests/e2e/sidecar ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/sidecar ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + jaeger_service_name=order + start_test sidecar-deployment + '[' 1 -ne 1 ']' + test_name=sidecar-deployment + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-deployment' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-deployment\e[0m' Rendering files for test sidecar-deployment + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build + '[' _build '!=' _build ']' + mkdir -p sidecar-deployment + cd sidecar-deployment + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-namespace + '[' 1 -ne 1 ']' + test_name=sidecar-namespace + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-namespace' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-namespace\e[0m' Rendering files for test sidecar-namespace + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-deployment + '[' sidecar-deployment '!=' _build ']' + cd .. + mkdir -p sidecar-namespace + cd sidecar-namespace + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-skip-webhook + '[' 1 -ne 1 ']' + test_name=sidecar-skip-webhook + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-skip-webhook' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-skip-webhook\e[0m' Rendering files for test sidecar-skip-webhook + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-namespace + '[' sidecar-namespace '!=' _build ']' + cd .. + mkdir -p sidecar-skip-webhook + cd sidecar-skip-webhook + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running sidecar E2E tests' Running sidecar E2E tests + cd tests/e2e/sidecar/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2576283944 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/sidecar-deployment === PAUSE kuttl/harness/sidecar-deployment === RUN kuttl/harness/sidecar-namespace === PAUSE kuttl/harness/sidecar-namespace === RUN kuttl/harness/sidecar-skip-webhook === PAUSE kuttl/harness/sidecar-skip-webhook === CONT kuttl/harness/artifacts logger.go:42: 07:32:32 | artifacts | Creating namespace: kuttl-test-adequate-egret logger.go:42: 07:32:32 | artifacts | artifacts events from ns kuttl-test-adequate-egret: logger.go:42: 07:32:32 | artifacts | Deleting namespace: kuttl-test-adequate-egret === CONT kuttl/harness/sidecar-namespace logger.go:42: 07:32:38 | sidecar-namespace | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:32:38 | sidecar-namespace | Creating namespace: kuttl-test-real-porpoise logger.go:42: 07:32:38 | sidecar-namespace/0-install | starting test step 0-install logger.go:42: 07:32:38 | sidecar-namespace/0-install | Jaeger:kuttl-test-real-porpoise/agent-as-sidecar created logger.go:42: 07:32:44 | sidecar-namespace/0-install | test step completed 0-install logger.go:42: 07:32:44 | sidecar-namespace/1-install | starting test step 1-install logger.go:42: 07:32:44 | sidecar-namespace/1-install | Deployment:kuttl-test-real-porpoise/vertx-create-span-sidecar created logger.go:42: 07:32:50 | sidecar-namespace/1-install | test step completed 1-install logger.go:42: 07:32:50 | sidecar-namespace/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:32:50 | sidecar-namespace/2-enable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="true"] logger.go:42: 07:32:51 | sidecar-namespace/2-enable-injection | namespace/kuttl-test-real-porpoise annotate logger.go:42: 07:32:53 | sidecar-namespace/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:32:53 | sidecar-namespace/3-find-service | starting test step 3-find-service logger.go:42: 07:32:53 | sidecar-namespace/3-find-service | Job:kuttl-test-real-porpoise/00-find-service created logger.go:42: 07:33:06 | sidecar-namespace/3-find-service | test step completed 3-find-service logger.go:42: 07:33:06 | sidecar-namespace/4-other-instance | starting test step 4-other-instance logger.go:42: 07:33:06 | sidecar-namespace/4-other-instance | Jaeger:kuttl-test-real-porpoise/agent-as-sidecar2 created logger.go:42: 07:33:14 | sidecar-namespace/4-other-instance | test step completed 4-other-instance logger.go:42: 07:33:14 | sidecar-namespace/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:33:15 | sidecar-namespace/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:33:15 | sidecar-namespace/6-find-service | starting test step 6-find-service logger.go:42: 07:33:15 | sidecar-namespace/6-find-service | Job:kuttl-test-real-porpoise/01-find-service created logger.go:42: 07:33:35 | sidecar-namespace/6-find-service | test step completed 6-find-service logger.go:42: 07:33:35 | sidecar-namespace/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:33:35 | sidecar-namespace/7-disable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="false"] logger.go:42: 07:33:35 | sidecar-namespace/7-disable-injection | namespace/kuttl-test-real-porpoise annotate logger.go:42: 07:33:37 | sidecar-namespace/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:33:37 | sidecar-namespace | sidecar-namespace events from ns kuttl-test-real-porpoise: logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:42 +0000 UTC Normal Pod agent-as-sidecar-6d59896bf5-dgmqg Binding Scheduled Successfully assigned kuttl-test-real-porpoise/agent-as-sidecar-6d59896bf5-dgmqg to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:42 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-6d59896bf5 SuccessfulCreate Created pod: agent-as-sidecar-6d59896bf5-dgmqg replicaset-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:42 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-6d59896bf5 to 1 deployment-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:43 +0000 UTC Normal Pod agent-as-sidecar-6d59896bf5-dgmqg AddedInterface Add eth0 [10.131.0.65/23] from ovn-kubernetes logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:43 +0000 UTC Normal Pod agent-as-sidecar-6d59896bf5-dgmqg.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:43 +0000 UTC Normal Pod agent-as-sidecar-6d59896bf5-dgmqg.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:43 +0000 UTC Normal Pod agent-as-sidecar-6d59896bf5-dgmqg.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:44 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-kmjjn Binding Scheduled Successfully assigned kuttl-test-real-porpoise/vertx-create-span-sidecar-84d458b68c-kmjjn to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:44 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-kmjjn replicaset-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:44 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:45 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-kmjjn AddedInterface Add eth0 [10.128.2.93/23] from ovn-kubernetes logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:45 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-kmjjn.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:49 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-kmjjn.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.652s (3.652s including waiting) kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:49 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-kmjjn.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:49 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-kmjjn.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:51 +0000 UTC Normal Pod vertx-create-span-sidecar-59555764f9-twxzl Binding Scheduled Successfully assigned kuttl-test-real-porpoise/vertx-create-span-sidecar-59555764f9-twxzl to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:51 +0000 UTC Normal Pod vertx-create-span-sidecar-59555764f9-twxzl AddedInterface Add eth0 [10.129.2.54/23] from ovn-kubernetes logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:51 +0000 UTC Normal Pod vertx-create-span-sidecar-59555764f9-twxzl.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:51 +0000 UTC Normal Pod vertx-create-span-sidecar-59555764f9-twxzl.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:51 +0000 UTC Normal Pod vertx-create-span-sidecar-59555764f9-twxzl.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:51 +0000 UTC Normal Pod vertx-create-span-sidecar-59555764f9-twxzl.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:51 +0000 UTC Normal Pod vertx-create-span-sidecar-59555764f9-twxzl.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:51 +0000 UTC Normal Pod vertx-create-span-sidecar-59555764f9-twxzl.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:51 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-59555764f9 SuccessfulCreate Created pod: vertx-create-span-sidecar-59555764f9-twxzl replicaset-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:51 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-59555764f9 to 1 deployment-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:53 +0000 UTC Normal Pod 00-find-service-9ncxk Binding Scheduled Successfully assigned kuttl-test-real-porpoise/00-find-service-9ncxk to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:53 +0000 UTC Normal Pod 00-find-service-9ncxk AddedInterface Add eth0 [10.131.0.66/23] from ovn-kubernetes logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:53 +0000 UTC Normal Pod 00-find-service-9ncxk.spec.containers{asserts-container} Pulling Pulling image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:53 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-9ncxk job-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:55 +0000 UTC Normal Pod 00-find-service-9ncxk.spec.containers{asserts-container} Pulled Successfully pulled image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" in 1.423s (1.423s including waiting) kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:55 +0000 UTC Normal Pod 00-find-service-9ncxk.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:55 +0000 UTC Normal Pod 00-find-service-9ncxk.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:57 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-kmjjn.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.93:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:57 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-kmjjn.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.93:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:59 +0000 UTC Warning Pod vertx-create-span-sidecar-59555764f9-twxzl.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.54:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:59 +0000 UTC Warning Pod vertx-create-span-sidecar-59555764f9-twxzl.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.54:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:32:59 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-kmjjn.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:00 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-kmjjn.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.93:8080/": read tcp 10.128.2.2:50548->10.128.2.93:8080: read: connection reset by peer kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:00 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-kmjjn.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.93:8080/": dial tcp 10.128.2.93:8080: connect: connection refused kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:00 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-kmjjn.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:01 +0000 UTC Normal Pod vertx-create-span-sidecar-59555764f9-twxzl.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:02 +0000 UTC Warning Pod vertx-create-span-sidecar-59555764f9-twxzl.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.54:8080/": read tcp 10.129.2.2:35044->10.129.2.54:8080: read: connection reset by peer kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:02 +0000 UTC Warning Pod vertx-create-span-sidecar-59555764f9-twxzl.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.54:8080/": dial tcp 10.129.2.54:8080: connect: connection refused kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:06 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:11 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-kmjjn.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.93:8080/": read tcp 10.128.2.2:46238->10.128.2.93:8080: read: connection reset by peer kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:12 +0000 UTC Normal Pod agent-as-sidecar2-7947cdc745-cxxfb Binding Scheduled Successfully assigned kuttl-test-real-porpoise/agent-as-sidecar2-7947cdc745-cxxfb to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:12 +0000 UTC Warning Pod agent-as-sidecar2-7947cdc745-cxxfb FailedMount MountVolume.SetUp failed for volume "agent-as-sidecar2-collector-tls-config-volume" : secret "agent-as-sidecar2-collector-headless-tls" not found kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:12 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-7947cdc745 SuccessfulCreate Created pod: agent-as-sidecar2-7947cdc745-cxxfb replicaset-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:12 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-7947cdc745 to 1 deployment-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:13 +0000 UTC Normal Pod agent-as-sidecar2-7947cdc745-cxxfb AddedInterface Add eth0 [10.131.0.67/23] from ovn-kubernetes logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:13 +0000 UTC Normal Pod agent-as-sidecar2-7947cdc745-cxxfb.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:13 +0000 UTC Normal Pod agent-as-sidecar2-7947cdc745-cxxfb.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:13 +0000 UTC Normal Pod agent-as-sidecar2-7947cdc745-cxxfb.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:14 +0000 UTC Normal Pod agent-as-sidecar-6d59896bf5-dgmqg.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:15 +0000 UTC Normal Pod 01-find-service-kk2bh Binding Scheduled Successfully assigned kuttl-test-real-porpoise/01-find-service-kk2bh to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:15 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-kk2bh job-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:16 +0000 UTC Normal Pod 01-find-service-kk2bh AddedInterface Add eth0 [10.131.0.68/23] from ovn-kubernetes logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:16 +0000 UTC Normal Pod 01-find-service-kk2bh.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:16 +0000 UTC Normal Pod 01-find-service-kk2bh.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:16 +0000 UTC Normal Pod 01-find-service-kk2bh.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:17 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-kmjjn replicaset-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:17 +0000 UTC Normal Pod vertx-create-span-sidecar-97b8ddff7-9h55z Binding Scheduled Successfully assigned kuttl-test-real-porpoise/vertx-create-span-sidecar-97b8ddff7-9h55z to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:17 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-97b8ddff7 SuccessfulCreate Created pod: vertx-create-span-sidecar-97b8ddff7-9h55z replicaset-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:17 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:17 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-97b8ddff7 to 1 from 0 deployment-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:18 +0000 UTC Normal Pod vertx-create-span-sidecar-97b8ddff7-9h55z AddedInterface Add eth0 [10.128.2.94/23] from ovn-kubernetes logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:18 +0000 UTC Normal Pod vertx-create-span-sidecar-97b8ddff7-9h55z.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:18 +0000 UTC Normal Pod vertx-create-span-sidecar-97b8ddff7-9h55z.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:18 +0000 UTC Normal Pod vertx-create-span-sidecar-97b8ddff7-9h55z.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:18 +0000 UTC Normal Pod vertx-create-span-sidecar-97b8ddff7-9h55z.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:18 +0000 UTC Normal Pod vertx-create-span-sidecar-97b8ddff7-9h55z.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:18 +0000 UTC Normal Pod vertx-create-span-sidecar-97b8ddff7-9h55z.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:26 +0000 UTC Warning Pod vertx-create-span-sidecar-97b8ddff7-9h55z.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.94:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:26 +0000 UTC Warning Pod vertx-create-span-sidecar-97b8ddff7-9h55z.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.94:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:29 +0000 UTC Normal Pod vertx-create-span-sidecar-97b8ddff7-9h55z.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:29 +0000 UTC Warning Pod vertx-create-span-sidecar-97b8ddff7-9h55z.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.94:8080/": read tcp 10.128.2.2:39842->10.128.2.94:8080: read: connection reset by peer kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:29 +0000 UTC Warning Pod vertx-create-span-sidecar-97b8ddff7-9h55z.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.94:8080/": dial tcp 10.128.2.94:8080: connect: connection refused kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:35 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:35 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-59555764f9 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-59555764f9-twxzl replicaset-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:35 +0000 UTC Normal Pod vertx-create-span-sidecar-b44d959bd-tdxqx Binding Scheduled Successfully assigned kuttl-test-real-porpoise/vertx-create-span-sidecar-b44d959bd-tdxqx to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:35 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-b44d959bd SuccessfulCreate Created pod: vertx-create-span-sidecar-b44d959bd-tdxqx replicaset-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:35 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-59555764f9 to 0 from 1 deployment-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:35 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-b44d959bd to 1 from 0 deployment-controller logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:36 +0000 UTC Normal Pod vertx-create-span-sidecar-b44d959bd-tdxqx AddedInterface Add eth0 [10.131.0.69/23] from ovn-kubernetes logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:36 +0000 UTC Normal Pod vertx-create-span-sidecar-b44d959bd-tdxqx.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:36 +0000 UTC Normal Pod vertx-create-span-sidecar-b44d959bd-tdxqx.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:33:37 | sidecar-namespace | 2024-01-22 07:33:36 +0000 UTC Normal Pod vertx-create-span-sidecar-b44d959bd-tdxqx.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:33:37 | sidecar-namespace | Deleting namespace: kuttl-test-real-porpoise === CONT kuttl/harness/sidecar-skip-webhook logger.go:42: 07:33:44 | sidecar-skip-webhook | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:33:44 | sidecar-skip-webhook | Creating namespace: kuttl-test-immortal-griffon logger.go:42: 07:33:44 | sidecar-skip-webhook/0-install | starting test step 0-install logger.go:42: 07:33:44 | sidecar-skip-webhook/0-install | Jaeger:kuttl-test-immortal-griffon/agent-as-sidecar created logger.go:42: 07:33:50 | sidecar-skip-webhook/0-install | test step completed 0-install logger.go:42: 07:33:50 | sidecar-skip-webhook/1-install | starting test step 1-install logger.go:42: 07:33:50 | sidecar-skip-webhook/1-install | Deployment:kuttl-test-immortal-griffon/vertx-create-span-sidecar created logger.go:42: 07:33:51 | sidecar-skip-webhook/1-install | test step completed 1-install logger.go:42: 07:33:51 | sidecar-skip-webhook/2-add-anotation-and-label | starting test step 2-add-anotation-and-label logger.go:42: 07:33:51 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name=jaeger-operator --namespace kuttl-test-immortal-griffon] logger.go:42: 07:33:51 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar labeled logger.go:42: 07:33:51 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-immortal-griffon] logger.go:42: 07:33:51 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar annotate logger.go:42: 07:33:51 | sidecar-skip-webhook/2-add-anotation-and-label | test step completed 2-add-anotation-and-label logger.go:42: 07:33:51 | sidecar-skip-webhook/3-remove-label | starting test step 3-remove-label logger.go:42: 07:33:51 | sidecar-skip-webhook/3-remove-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name- --namespace kuttl-test-immortal-griffon] logger.go:42: 07:33:51 | sidecar-skip-webhook/3-remove-label | deployment.apps/vertx-create-span-sidecar unlabeled logger.go:42: 07:33:54 | sidecar-skip-webhook/3-remove-label | test step completed 3-remove-label logger.go:42: 07:33:54 | sidecar-skip-webhook | sidecar-skip-webhook events from ns kuttl-test-immortal-griffon: logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:48 +0000 UTC Normal Pod agent-as-sidecar-c5c5579d8-zvkrm Binding Scheduled Successfully assigned kuttl-test-immortal-griffon/agent-as-sidecar-c5c5579d8-zvkrm to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:48 +0000 UTC Normal Pod agent-as-sidecar-c5c5579d8-zvkrm AddedInterface Add eth0 [10.131.0.70/23] from ovn-kubernetes logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:48 +0000 UTC Normal Pod agent-as-sidecar-c5c5579d8-zvkrm.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:48 +0000 UTC Normal Pod agent-as-sidecar-c5c5579d8-zvkrm.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:48 +0000 UTC Normal Pod agent-as-sidecar-c5c5579d8-zvkrm.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:48 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-c5c5579d8 SuccessfulCreate Created pod: agent-as-sidecar-c5c5579d8-zvkrm replicaset-controller logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:48 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-c5c5579d8 to 1 deployment-controller logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:50 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-xzk8c Binding Scheduled Successfully assigned kuttl-test-immortal-griffon/vertx-create-span-sidecar-84d458b68c-xzk8c to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:50 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-xzk8c replicaset-controller logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:50 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:51 +0000 UTC Normal Pod vertx-create-span-sidecar-75fcc44954-phvtn Binding Scheduled Successfully assigned kuttl-test-immortal-griffon/vertx-create-span-sidecar-75fcc44954-phvtn to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:51 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-75fcc44954 SuccessfulCreate Created pod: vertx-create-span-sidecar-75fcc44954-phvtn replicaset-controller logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:51 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-xzk8c AddedInterface Add eth0 [10.128.2.95/23] from ovn-kubernetes logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:51 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-xzk8c.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:51 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-xzk8c.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:51 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-xzk8c.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:51 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-75fcc44954 to 1 deployment-controller logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:52 +0000 UTC Normal Pod vertx-create-span-sidecar-75fcc44954-phvtn AddedInterface Add eth0 [10.131.0.71/23] from ovn-kubernetes logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:52 +0000 UTC Normal Pod vertx-create-span-sidecar-75fcc44954-phvtn.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:52 +0000 UTC Normal Pod vertx-create-span-sidecar-75fcc44954-phvtn.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:52 +0000 UTC Normal Pod vertx-create-span-sidecar-75fcc44954-phvtn.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:52 +0000 UTC Normal Pod vertx-create-span-sidecar-75fcc44954-phvtn.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:52 +0000 UTC Normal Pod vertx-create-span-sidecar-75fcc44954-phvtn.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:33:54 | sidecar-skip-webhook | 2024-01-22 07:33:52 +0000 UTC Normal Pod vertx-create-span-sidecar-75fcc44954-phvtn.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:33:54 | sidecar-skip-webhook | Deleting namespace: kuttl-test-immortal-griffon === CONT kuttl/harness/sidecar-deployment logger.go:42: 07:34:00 | sidecar-deployment | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:34:00 | sidecar-deployment | Creating namespace: kuttl-test-capable-octopus logger.go:42: 07:34:00 | sidecar-deployment/0-install | starting test step 0-install logger.go:42: 07:34:00 | sidecar-deployment/0-install | Jaeger:kuttl-test-capable-octopus/agent-as-sidecar created logger.go:42: 07:34:06 | sidecar-deployment/0-install | test step completed 0-install logger.go:42: 07:34:06 | sidecar-deployment/1-install | starting test step 1-install logger.go:42: 07:34:06 | sidecar-deployment/1-install | Deployment:kuttl-test-capable-octopus/vertx-create-span-sidecar created logger.go:42: 07:34:08 | sidecar-deployment/1-install | test step completed 1-install logger.go:42: 07:34:08 | sidecar-deployment/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:34:08 | sidecar-deployment/2-enable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-capable-octopus] logger.go:42: 07:34:08 | sidecar-deployment/2-enable-injection | deployment.apps/vertx-create-span-sidecar annotate logger.go:42: 07:34:09 | sidecar-deployment/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:34:09 | sidecar-deployment/3-find-service | starting test step 3-find-service logger.go:42: 07:34:09 | sidecar-deployment/3-find-service | Job:kuttl-test-capable-octopus/00-find-service created logger.go:42: 07:34:21 | sidecar-deployment/3-find-service | test step completed 3-find-service logger.go:42: 07:34:21 | sidecar-deployment/4-other-instance | starting test step 4-other-instance logger.go:42: 07:34:21 | sidecar-deployment/4-other-instance | Jaeger:kuttl-test-capable-octopus/agent-as-sidecar2 created logger.go:42: 07:34:29 | sidecar-deployment/4-other-instance | test step completed 4-other-instance logger.go:42: 07:34:29 | sidecar-deployment/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:34:30 | sidecar-deployment/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:34:30 | sidecar-deployment/6-find-service | starting test step 6-find-service logger.go:42: 07:34:30 | sidecar-deployment/6-find-service | Job:kuttl-test-capable-octopus/01-find-service created logger.go:42: 07:34:49 | sidecar-deployment/6-find-service | test step completed 6-find-service logger.go:42: 07:34:49 | sidecar-deployment/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:34:49 | sidecar-deployment/7-disable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=false --namespace kuttl-test-capable-octopus] logger.go:42: 07:34:49 | sidecar-deployment/7-disable-injection | deployment.apps/vertx-create-span-sidecar annotate logger.go:42: 07:34:52 | sidecar-deployment/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:34:52 | sidecar-deployment | sidecar-deployment events from ns kuttl-test-capable-octopus: logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:03 +0000 UTC Normal Pod agent-as-sidecar-5c5869998c-tkpr5 Binding Scheduled Successfully assigned kuttl-test-capable-octopus/agent-as-sidecar-5c5869998c-tkpr5 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:03 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-5c5869998c SuccessfulCreate Created pod: agent-as-sidecar-5c5869998c-tkpr5 replicaset-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:03 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-5c5869998c to 1 deployment-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:04 +0000 UTC Normal Pod agent-as-sidecar-5c5869998c-tkpr5 AddedInterface Add eth0 [10.131.0.72/23] from ovn-kubernetes logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:04 +0000 UTC Normal Pod agent-as-sidecar-5c5869998c-tkpr5.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:04 +0000 UTC Normal Pod agent-as-sidecar-5c5869998c-tkpr5.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:04 +0000 UTC Normal Pod agent-as-sidecar-5c5869998c-tkpr5.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:06 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-d4hvc Binding Scheduled Successfully assigned kuttl-test-capable-octopus/vertx-create-span-sidecar-84d458b68c-d4hvc to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:06 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-d4hvc AddedInterface Add eth0 [10.128.2.96/23] from ovn-kubernetes logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:06 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-d4hvc.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:06 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-d4hvc.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:06 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-d4hvc replicaset-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:06 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:07 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-d4hvc.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:08 +0000 UTC Normal Pod vertx-create-span-sidecar-57d4b78547-l749t Binding Scheduled Successfully assigned kuttl-test-capable-octopus/vertx-create-span-sidecar-57d4b78547-l749t to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:08 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-57d4b78547 SuccessfulCreate Created pod: vertx-create-span-sidecar-57d4b78547-l749t replicaset-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:08 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-57d4b78547 to 1 deployment-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:09 +0000 UTC Normal Pod 00-find-service-6vrkg Binding Scheduled Successfully assigned kuttl-test-capable-octopus/00-find-service-6vrkg to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:09 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-6vrkg job-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:09 +0000 UTC Normal Pod vertx-create-span-sidecar-57d4b78547-l749t AddedInterface Add eth0 [10.131.0.73/23] from ovn-kubernetes logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:09 +0000 UTC Normal Pod vertx-create-span-sidecar-57d4b78547-l749t.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:09 +0000 UTC Normal Pod vertx-create-span-sidecar-57d4b78547-l749t.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:09 +0000 UTC Normal Pod vertx-create-span-sidecar-57d4b78547-l749t.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:09 +0000 UTC Normal Pod vertx-create-span-sidecar-57d4b78547-l749t.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:09 +0000 UTC Normal Pod vertx-create-span-sidecar-57d4b78547-l749t.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:09 +0000 UTC Normal Pod vertx-create-span-sidecar-57d4b78547-l749t.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:10 +0000 UTC Normal Pod 00-find-service-6vrkg AddedInterface Add eth0 [10.129.2.55/23] from ovn-kubernetes logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:10 +0000 UTC Normal Pod 00-find-service-6vrkg.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:10 +0000 UTC Normal Pod 00-find-service-6vrkg.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:10 +0000 UTC Normal Pod 00-find-service-6vrkg.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:14 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-d4hvc.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.96:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:14 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-d4hvc.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.96:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:16 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-d4hvc.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:17 +0000 UTC Warning Pod vertx-create-span-sidecar-57d4b78547-l749t.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.73:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:17 +0000 UTC Warning Pod vertx-create-span-sidecar-57d4b78547-l749t.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.73:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:17 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-d4hvc.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.96:8080/": read tcp 10.128.2.2:33928->10.128.2.96:8080: read: connection reset by peer kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:17 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-d4hvc.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.96:8080/": dial tcp 10.128.2.96:8080: connect: connection refused kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:19 +0000 UTC Normal Pod vertx-create-span-sidecar-57d4b78547-l749t.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:20 +0000 UTC Warning Pod vertx-create-span-sidecar-57d4b78547-l749t.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.73:8080/": read tcp 10.131.0.2:38640->10.131.0.73:8080: read: connection reset by peer kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:20 +0000 UTC Warning Pod vertx-create-span-sidecar-57d4b78547-l749t.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.73:8080/": dial tcp 10.131.0.73:8080: connect: connection refused kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:21 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:25 +0000 UTC Normal Pod agent-as-sidecar2-6548bf7d77-zksx9 Binding Scheduled Successfully assigned kuttl-test-capable-octopus/agent-as-sidecar2-6548bf7d77-zksx9 to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:25 +0000 UTC Normal Pod agent-as-sidecar2-6548bf7d77-zksx9 AddedInterface Add eth0 [10.129.2.56/23] from ovn-kubernetes logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:25 +0000 UTC Normal Pod agent-as-sidecar2-6548bf7d77-zksx9.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:25 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-6548bf7d77 SuccessfulCreate Created pod: agent-as-sidecar2-6548bf7d77-zksx9 replicaset-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:25 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-6548bf7d77 to 1 deployment-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:28 +0000 UTC Normal Pod agent-as-sidecar2-6548bf7d77-zksx9.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" in 2.967s (2.967s including waiting) kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:28 +0000 UTC Normal Pod agent-as-sidecar2-6548bf7d77-zksx9.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:28 +0000 UTC Normal Pod agent-as-sidecar2-6548bf7d77-zksx9.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:28 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-d4hvc.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.96:8080/": read tcp 10.128.2.2:60996->10.128.2.96:8080: read: connection reset by peer kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:29 +0000 UTC Normal Pod agent-as-sidecar-5c5869998c-tkpr5.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:30 +0000 UTC Normal Pod 01-find-service-wkw6r Binding Scheduled Successfully assigned kuttl-test-capable-octopus/01-find-service-wkw6r to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:30 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-wkw6r job-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:31 +0000 UTC Normal Pod 01-find-service-wkw6r AddedInterface Add eth0 [10.128.2.97/23] from ovn-kubernetes logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:31 +0000 UTC Normal Pod 01-find-service-wkw6r.spec.containers{asserts-container} Pulled Container image "registry.build03.ci.openshift.org/ci-op-zz0glpxl/pipeline@sha256:cd567af433d8e227d1cc8b5eabf6110c00cd782857f5db0e4d3962128793c882" already present on machine kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:31 +0000 UTC Normal Pod 01-find-service-wkw6r.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:31 +0000 UTC Normal Pod 01-find-service-wkw6r.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:32 +0000 UTC Normal Pod vertx-create-span-sidecar-64dd7dc46f-tnb6h Binding Scheduled Successfully assigned kuttl-test-capable-octopus/vertx-create-span-sidecar-64dd7dc46f-tnb6h to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:32 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-64dd7dc46f SuccessfulCreate Created pod: vertx-create-span-sidecar-64dd7dc46f-tnb6h replicaset-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:32 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-d4hvc replicaset-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:32 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:32 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-64dd7dc46f to 1 from 0 deployment-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:33 +0000 UTC Normal Pod vertx-create-span-sidecar-64dd7dc46f-tnb6h AddedInterface Add eth0 [10.129.2.57/23] from ovn-kubernetes logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:33 +0000 UTC Normal Pod vertx-create-span-sidecar-64dd7dc46f-tnb6h.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:33 +0000 UTC Normal Pod vertx-create-span-sidecar-64dd7dc46f-tnb6h.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:33 +0000 UTC Normal Pod vertx-create-span-sidecar-64dd7dc46f-tnb6h.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:33 +0000 UTC Normal Pod vertx-create-span-sidecar-64dd7dc46f-tnb6h.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:33 +0000 UTC Normal Pod vertx-create-span-sidecar-64dd7dc46f-tnb6h.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:33 +0000 UTC Normal Pod vertx-create-span-sidecar-64dd7dc46f-tnb6h.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:41 +0000 UTC Warning Pod vertx-create-span-sidecar-64dd7dc46f-tnb6h.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.57:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:42 +0000 UTC Warning Pod vertx-create-span-sidecar-64dd7dc46f-tnb6h.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.57:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:44 +0000 UTC Normal Pod vertx-create-span-sidecar-64dd7dc46f-tnb6h.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:44 +0000 UTC Warning Pod vertx-create-span-sidecar-64dd7dc46f-tnb6h.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.57:8080/": read tcp 10.129.2.2:43232->10.129.2.57:8080: read: connection reset by peer kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:44 +0000 UTC Warning Pod vertx-create-span-sidecar-64dd7dc46f-tnb6h.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.57:8080/": dial tcp 10.129.2.57:8080: connect: connection refused kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:49 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:49 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-57d4b78547 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-57d4b78547-l749t replicaset-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:49 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-57d4b78547 to 0 from 1 deployment-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:49 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-5ff9cdd496 to 1 from 0 deployment-controller logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:50 +0000 UTC Normal Pod vertx-create-span-sidecar-5ff9cdd496-wj8nv Binding Scheduled Successfully assigned kuttl-test-capable-octopus/vertx-create-span-sidecar-5ff9cdd496-wj8nv to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:50 +0000 UTC Normal Pod vertx-create-span-sidecar-5ff9cdd496-wj8nv AddedInterface Add eth0 [10.128.2.98/23] from ovn-kubernetes logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:50 +0000 UTC Normal Pod vertx-create-span-sidecar-5ff9cdd496-wj8nv.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:50 +0000 UTC Normal Pod vertx-create-span-sidecar-5ff9cdd496-wj8nv.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:50 +0000 UTC Normal Pod vertx-create-span-sidecar-5ff9cdd496-wj8nv.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:34:52 | sidecar-deployment | 2024-01-22 07:34:50 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-5ff9cdd496 SuccessfulCreate Created pod: vertx-create-span-sidecar-5ff9cdd496-wj8nv replicaset-controller logger.go:42: 07:34:52 | sidecar-deployment | Deleting namespace: kuttl-test-capable-octopus === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (146.27s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.85s) --- PASS: kuttl/harness/sidecar-namespace (65.93s) --- PASS: kuttl/harness/sidecar-skip-webhook (15.68s) --- PASS: kuttl/harness/sidecar-deployment (58.77s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name sidecar --report --output /logs/artifacts/sidecar.xml ./artifacts/kuttl-report.xml time="2024-01-22T07:34:59Z" level=debug msg="Setting a new name for the test suites" time="2024-01-22T07:34:59Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-01-22T07:34:59Z" level=debug msg="normalizing test case names" time="2024-01-22T07:34:59Z" level=debug msg="sidecar/artifacts -> sidecar_artifacts" time="2024-01-22T07:34:59Z" level=debug msg="sidecar/sidecar-namespace -> sidecar_sidecar_namespace" time="2024-01-22T07:34:59Z" level=debug msg="sidecar/sidecar-skip-webhook -> sidecar_sidecar_skip_webhook" time="2024-01-22T07:34:59Z" level=debug msg="sidecar/sidecar-deployment -> sidecar_sidecar_deployment" +------------------------------+--------+ | NAME | RESULT | +------------------------------+--------+ | sidecar_artifacts | passed | | sidecar_sidecar_namespace | passed | | sidecar_sidecar_skip_webhook | passed | | sidecar_sidecar_deployment | passed | +------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh streaming false true + '[' 3 -ne 3 ']' + test_suite_name=streaming + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/streaming.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-streaming make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ SKIP_ES_EXTERNAL=true \ ./tests/e2e/streaming/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-18-050837 True False 56m Cluster version is 4.15.0-0.nightly-2024-01-18-050837' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-18-050837 True False 56m Cluster version is 4.15.0-0.nightly-2024-01-18-050837' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/streaming/render.sh ++ export SUITE_DIR=./tests/e2e/streaming ++ SUITE_DIR=./tests/e2e/streaming ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/streaming ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + '[' false = true ']' + start_test streaming-simple + '[' 1 -ne 1 ']' + test_name=streaming-simple + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-simple' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-simple\e[0m' Rendering files for test streaming-simple + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + mkdir -p streaming-simple + cd streaming-simple + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/streaming-jaeger-assert.yaml.template -o ./04-assert.yaml + render_smoke_test simple-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=simple-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + export JAEGER_NAME=simple-streaming + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-tls + '[' 1 -ne 1 ']' + test_name=streaming-with-tls + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-tls' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-tls\e[0m' Rendering files for test streaming-with-tls + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-simple + '[' streaming-simple '!=' _build ']' + cd .. + mkdir -p streaming-with-tls + cd streaming-with-tls + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + render_smoke_test tls-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=tls-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + export JAEGER_NAME=tls-streaming + JAEGER_NAME=tls-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-autoprovisioning-autoscale + '[' 1 -ne 1 ']' + test_name=streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-autoprovisioning-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-autoprovisioning-autoscale\e[0m' Rendering files for test streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-tls + '[' streaming-with-tls '!=' _build ']' + cd .. + mkdir -p streaming-with-autoprovisioning-autoscale + cd streaming-with-autoprovisioning-autoscale + '[' true = true ']' + rm ./00-install.yaml ./00-assert.yaml + render_install_elasticsearch upstream 01 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=01 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./01-assert.yaml + jaeger_name=auto-provisioned + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="20Mi"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="500m"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.autoscale=true ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.minReplicas=1 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.maxReplicas=2 ./02-install.yaml + render_assert_kafka true auto-provisioned 03 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provisioned + test_step=03 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./03-assert.yaml ++ expr 03 + 1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./04-assert.yaml ++ expr 03 + 2 + CLUSTER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./05-assert.yaml + version_lt 1.27 1.23 ++ echo 1.27 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.27 '!=' 1.27 + rm ./08-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running streaming E2E tests' Running streaming E2E tests + cd tests/e2e/streaming/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2576283944 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/streaming-simple === PAUSE kuttl/harness/streaming-simple === RUN kuttl/harness/streaming-with-autoprovisioning-autoscale === PAUSE kuttl/harness/streaming-with-autoprovisioning-autoscale === RUN kuttl/harness/streaming-with-tls === PAUSE kuttl/harness/streaming-with-tls === CONT kuttl/harness/artifacts logger.go:42: 07:35:12 | artifacts | Creating namespace: kuttl-test-prepared-gopher logger.go:42: 07:35:12 | artifacts | artifacts events from ns kuttl-test-prepared-gopher: logger.go:42: 07:35:12 | artifacts | Deleting namespace: kuttl-test-prepared-gopher === CONT kuttl/harness/streaming-with-autoprovisioning-autoscale logger.go:42: 07:35:19 | streaming-with-autoprovisioning-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:35:19 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:35:19 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:35:19 | streaming-with-autoprovisioning-autoscale | Creating namespace: kuttl-test-chief-crow logger.go:42: 07:35:19 | streaming-with-autoprovisioning-autoscale/1-install | starting test step 1-install logger.go:42: 07:35:19 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc create sa deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:35:19 | streaming-with-autoprovisioning-autoscale/1-install | serviceaccount/deploy-elasticsearch created logger.go:42: 07:35:19 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc adm policy add-scc-to-user privileged -z deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:35:19 | streaming-with-autoprovisioning-autoscale/1-install | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:privileged added: "deploy-elasticsearch" logger.go:42: 07:35:19 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 6] logger.go:42: 07:35:25 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_0.yml -n $NAMESPACE] logger.go:42: 07:35:25 | streaming-with-autoprovisioning-autoscale/1-install | statefulset.apps/elasticsearch created logger.go:42: 07:35:25 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 3] logger.go:42: 07:35:28 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_1.yml -n $NAMESPACE] logger.go:42: 07:35:29 | streaming-with-autoprovisioning-autoscale/1-install | service/elasticsearch created logger.go:42: 07:35:46 | streaming-with-autoprovisioning-autoscale/1-install | test step completed 1-install logger.go:42: 07:35:46 | streaming-with-autoprovisioning-autoscale/2-install | starting test step 2-install logger.go:42: 07:35:46 | streaming-with-autoprovisioning-autoscale/2-install | Jaeger:kuttl-test-chief-crow/auto-provisioned created logger.go:42: 07:35:46 | streaming-with-autoprovisioning-autoscale/2-install | test step completed 2-install logger.go:42: 07:35:46 | streaming-with-autoprovisioning-autoscale/3- | starting test step 3- logger.go:42: 07:36:37 | streaming-with-autoprovisioning-autoscale/3- | test step completed 3- logger.go:42: 07:36:37 | streaming-with-autoprovisioning-autoscale/4- | starting test step 4- logger.go:42: 07:37:09 | streaming-with-autoprovisioning-autoscale/4- | test step completed 4- logger.go:42: 07:37:09 | streaming-with-autoprovisioning-autoscale/5- | starting test step 5- logger.go:42: 07:37:42 | streaming-with-autoprovisioning-autoscale/5- | test step completed 5- logger.go:42: 07:37:42 | streaming-with-autoprovisioning-autoscale/6- | starting test step 6- logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale/6- | test step completed 6- logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale/7- | starting test step 7- logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale/7- | test step completed 7- logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | streaming-with-autoprovisioning-autoscale events from ns kuttl-test-chief-crow: logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:35:25 +0000 UTC Normal Pod elasticsearch-0 Binding Scheduled Successfully assigned kuttl-test-chief-crow/elasticsearch-0 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:35:25 +0000 UTC Normal StatefulSet.apps elasticsearch SuccessfulCreate create Pod elasticsearch-0 in StatefulSet elasticsearch successful statefulset-controller logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:35:26 +0000 UTC Normal Pod elasticsearch-0 AddedInterface Add eth0 [10.131.0.74/23] from ovn-kubernetes logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:35:26 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulling Pulling image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:35:35 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulled Successfully pulled image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" in 9.333s (9.333s including waiting) kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:35:35 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:35:35 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:35:41 +0000 UTC Warning Pod elasticsearch-0.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Get "http://10.131.0.74:9200/": dial tcp 10.131.0.74:9200: connect: connection refused kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:35:51 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-zookeeper NoPods No matching pods found controllermanager logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:35:51 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:35:51 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-chief-crow/data-auto-provisioned-zookeeper-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-749495bdfc-g7vwc_0df692b7-1769-48b2-a189-c6004da18fc6 logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:35:51 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:35:55 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ProvisioningSucceeded Successfully provisioned volume pvc-6ca8fc72-4394-419b-a5b2-b898d7bf2815 ebs.csi.aws.com_aws-ebs-csi-driver-controller-749495bdfc-g7vwc_0df692b7-1769-48b2-a189-c6004da18fc6 logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:35:56 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 Binding Scheduled Successfully assigned kuttl-test-chief-crow/auto-provisioned-zookeeper-0 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:35:58 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-6ca8fc72-4394-419b-a5b2-b898d7bf2815" attachdetach-controller logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:36:07 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 AddedInterface Add eth0 [10.128.2.99/23] from ovn-kubernetes logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:36:07 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:55b18000b2e774adccd0d5393b58fc3b8358ab174eaaa3ada9d520b5cde30669" kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:36:15 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:55b18000b2e774adccd0d5393b58fc3b8358ab174eaaa3ada9d520b5cde30669" in 7.827s (7.827s including waiting) kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:36:15 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Created Created container zookeeper kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:36:15 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Started Started container zookeeper kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:36:39 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-kafka NoPods No matching pods found controllermanager logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:36:39 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:36:39 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:36:39 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-chief-crow/data-0-auto-provisioned-kafka-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-749495bdfc-g7vwc_0df692b7-1769-48b2-a189-c6004da18fc6 logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:36:44 +0000 UTC Normal Pod auto-provisioned-kafka-0 Binding Scheduled Successfully assigned kuttl-test-chief-crow/auto-provisioned-kafka-0 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:36:44 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ProvisioningSucceeded Successfully provisioned volume pvc-a89e3752-9570-424e-9a95-f8d05133cfb7 ebs.csi.aws.com_aws-ebs-csi-driver-controller-749495bdfc-g7vwc_0df692b7-1769-48b2-a189-c6004da18fc6 logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:36:47 +0000 UTC Normal Pod auto-provisioned-kafka-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-a89e3752-9570-424e-9a95-f8d05133cfb7" attachdetach-controller logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:36:48 +0000 UTC Normal Pod auto-provisioned-kafka-0 AddedInterface Add eth0 [10.128.2.100/23] from ovn-kubernetes logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:36:48 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulled Container image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:55b18000b2e774adccd0d5393b58fc3b8358ab174eaaa3ada9d520b5cde30669" already present on machine kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:36:48 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Created Created container kafka kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:36:48 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Started Started container kafka kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:10 +0000 UTC Normal Deployment.apps auto-provisioned-entity-operator ScalingReplicaSet Scaled up replica set auto-provisioned-entity-operator-7694c7c9bd to 1 deployment-controller logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:11 +0000 UTC Normal Pod auto-provisioned-entity-operator-7694c7c9bd-xb56b Binding Scheduled Successfully assigned kuttl-test-chief-crow/auto-provisioned-entity-operator-7694c7c9bd-xb56b to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:11 +0000 UTC Normal Pod auto-provisioned-entity-operator-7694c7c9bd-xb56b AddedInterface Add eth0 [10.129.2.58/23] from ovn-kubernetes logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:11 +0000 UTC Normal Pod auto-provisioned-entity-operator-7694c7c9bd-xb56b.spec.containers{topic-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:f1be1aa2f18276f9169893eb55e3733cd52fa38f2101a9b3925f79774841689f" already present on machine kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:11 +0000 UTC Normal Pod auto-provisioned-entity-operator-7694c7c9bd-xb56b.spec.containers{topic-operator} Created Created container topic-operator kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:11 +0000 UTC Normal Pod auto-provisioned-entity-operator-7694c7c9bd-xb56b.spec.containers{topic-operator} Started Started container topic-operator kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:11 +0000 UTC Normal Pod auto-provisioned-entity-operator-7694c7c9bd-xb56b.spec.containers{user-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:f1be1aa2f18276f9169893eb55e3733cd52fa38f2101a9b3925f79774841689f" already present on machine kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:11 +0000 UTC Normal Pod auto-provisioned-entity-operator-7694c7c9bd-xb56b.spec.containers{user-operator} Created Created container user-operator kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:11 +0000 UTC Normal Pod auto-provisioned-entity-operator-7694c7c9bd-xb56b.spec.containers{user-operator} Started Started container user-operator kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:11 +0000 UTC Normal Pod auto-provisioned-entity-operator-7694c7c9bd-xb56b.spec.containers{tls-sidecar} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:55b18000b2e774adccd0d5393b58fc3b8358ab174eaaa3ada9d520b5cde30669" kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:11 +0000 UTC Normal ReplicaSet.apps auto-provisioned-entity-operator-7694c7c9bd SuccessfulCreate Created pod: auto-provisioned-entity-operator-7694c7c9bd-xb56b replicaset-controller logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:21 +0000 UTC Normal Pod auto-provisioned-entity-operator-7694c7c9bd-xb56b.spec.containers{tls-sidecar} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:55b18000b2e774adccd0d5393b58fc3b8358ab174eaaa3ada9d520b5cde30669" in 9.721s (9.721s including waiting) kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:21 +0000 UTC Normal Pod auto-provisioned-entity-operator-7694c7c9bd-xb56b.spec.containers{tls-sidecar} Created Created container tls-sidecar kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:21 +0000 UTC Normal Pod auto-provisioned-entity-operator-7694c7c9bd-xb56b.spec.containers{tls-sidecar} Started Started container tls-sidecar kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:44 +0000 UTC Normal Pod auto-provisioned-collector-54b564f8c-x5zm5 Binding Scheduled Successfully assigned kuttl-test-chief-crow/auto-provisioned-collector-54b564f8c-x5zm5 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:44 +0000 UTC Normal ReplicaSet.apps auto-provisioned-collector-54b564f8c SuccessfulCreate Created pod: auto-provisioned-collector-54b564f8c-x5zm5 replicaset-controller logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:44 +0000 UTC Normal Deployment.apps auto-provisioned-collector ScalingReplicaSet Scaled up replica set auto-provisioned-collector-54b564f8c to 1 deployment-controller logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:44 +0000 UTC Normal Pod auto-provisioned-ingester-6b7f7c6df7-jvsc5 Binding Scheduled Successfully assigned kuttl-test-chief-crow/auto-provisioned-ingester-6b7f7c6df7-jvsc5 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:44 +0000 UTC Normal ReplicaSet.apps auto-provisioned-ingester-6b7f7c6df7 SuccessfulCreate Created pod: auto-provisioned-ingester-6b7f7c6df7-jvsc5 replicaset-controller logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:44 +0000 UTC Normal Deployment.apps auto-provisioned-ingester ScalingReplicaSet Scaled up replica set auto-provisioned-ingester-6b7f7c6df7 to 1 deployment-controller logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:44 +0000 UTC Normal Pod auto-provisioned-query-68647c9648-ds25f Binding Scheduled Successfully assigned kuttl-test-chief-crow/auto-provisioned-query-68647c9648-ds25f to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:44 +0000 UTC Warning Pod auto-provisioned-query-68647c9648-ds25f FailedMount MountVolume.SetUp failed for volume "auto-provisioned-ui-oauth-proxy-tls" : secret "auto-provisioned-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:44 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-68647c9648 SuccessfulCreate Created pod: auto-provisioned-query-68647c9648-ds25f replicaset-controller logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:44 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled up replica set auto-provisioned-query-68647c9648 to 1 deployment-controller logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:45 +0000 UTC Warning Pod auto-provisioned-collector-54b564f8c-x5zm5 FailedMount MountVolume.SetUp failed for volume "kafkauser-auto-provisioned" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:45 +0000 UTC Warning Pod auto-provisioned-collector-54b564f8c-x5zm5 FailedMount MountVolume.SetUp failed for volume "auto-provisioned-sampling-configuration-volume" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:45 +0000 UTC Warning Pod auto-provisioned-collector-54b564f8c-x5zm5 FailedMount MountVolume.SetUp failed for volume "auto-provisioned-trusted-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:45 +0000 UTC Warning Pod auto-provisioned-ingester-6b7f7c6df7-jvsc5 FailedMount MountVolume.SetUp failed for volume "auto-provisioned-trusted-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:45 +0000 UTC Warning Pod auto-provisioned-query-68647c9648-ds25f FailedMount MountVolume.SetUp failed for volume "kafkauser-auto-provisioned" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:45 +0000 UTC Warning Pod auto-provisioned-query-68647c9648-ds25f FailedMount MountVolume.SetUp failed for volume "auto-provisioned-trusted-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:46 +0000 UTC Normal Pod auto-provisioned-collector-54b564f8c-x5zm5 AddedInterface Add eth0 [10.131.0.75/23] from ovn-kubernetes logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:46 +0000 UTC Normal Pod auto-provisioned-collector-54b564f8c-x5zm5.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:46 +0000 UTC Normal Pod auto-provisioned-collector-54b564f8c-x5zm5.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:46 +0000 UTC Normal Pod auto-provisioned-collector-54b564f8c-x5zm5.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:46 +0000 UTC Normal Pod auto-provisioned-ingester-6b7f7c6df7-jvsc5 AddedInterface Add eth0 [10.128.2.101/23] from ovn-kubernetes logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:46 +0000 UTC Normal Pod auto-provisioned-ingester-6b7f7c6df7-jvsc5.spec.containers{jaeger-ingester} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:849018528225b7370cc4740fc9f94bef7ffd4195328a916a6013d88f885eebe2" kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:46 +0000 UTC Normal Pod auto-provisioned-query-68647c9648-ds25f AddedInterface Add eth0 [10.131.0.76/23] from ovn-kubernetes logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:46 +0000 UTC Normal Pod auto-provisioned-query-68647c9648-ds25f.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:46 +0000 UTC Normal Pod auto-provisioned-query-68647c9648-ds25f.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:46 +0000 UTC Normal Pod auto-provisioned-query-68647c9648-ds25f.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:46 +0000 UTC Normal Pod auto-provisioned-query-68647c9648-ds25f.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:46 +0000 UTC Normal Pod auto-provisioned-query-68647c9648-ds25f.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:46 +0000 UTC Normal Pod auto-provisioned-query-68647c9648-ds25f.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:46 +0000 UTC Normal Pod auto-provisioned-query-68647c9648-ds25f.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:46 +0000 UTC Normal Pod auto-provisioned-query-68647c9648-ds25f.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:46 +0000 UTC Normal Pod auto-provisioned-query-68647c9648-ds25f.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:52 +0000 UTC Normal Pod auto-provisioned-ingester-6b7f7c6df7-jvsc5.spec.containers{jaeger-ingester} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:849018528225b7370cc4740fc9f94bef7ffd4195328a916a6013d88f885eebe2" in 5.828s (5.828s including waiting) kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:52 +0000 UTC Normal Pod auto-provisioned-ingester-6b7f7c6df7-jvsc5.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | 2024-01-22 07:37:52 +0000 UTC Normal Pod auto-provisioned-ingester-6b7f7c6df7-jvsc5.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 07:37:53 | streaming-with-autoprovisioning-autoscale | Deleting namespace: kuttl-test-chief-crow === CONT kuttl/harness/streaming-with-tls logger.go:42: 07:38:33 | streaming-with-tls | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:38:33 | streaming-with-tls | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:38:33 | streaming-with-tls | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:38:34 | streaming-with-tls | Creating namespace: kuttl-test-obliging-maggot logger.go:42: 07:38:34 | streaming-with-tls/0-install | starting test step 0-install logger.go:42: 07:38:34 | streaming-with-tls/0-install | running command: [sh -c cd /tmp/jaeger-tests && make undeploy-kafka KAFKA_NAMESPACE=$NAMESPACE KAFKA_OLM=true] logger.go:42: 07:38:34 | streaming-with-tls/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:38:34 | streaming-with-tls/0-install | >>>> Skiping kafka-operator undeploy logger.go:42: 07:38:34 | streaming-with-tls/0-install | kubectl delete --namespace kuttl-test-obliging-maggot -f tests/_build/kafka-example.yaml 2>&1 || true logger.go:42: 07:38:34 | streaming-with-tls/0-install | error: the path "tests/_build/kafka-example.yaml" does not exist logger.go:42: 07:38:34 | streaming-with-tls/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:38:34 | streaming-with-tls/0-install | running command: [sh -c cd /tmp/jaeger-tests && make kafka KAFKA_NAMESPACE=$NAMESPACE KAFKA_OLM=true KAFKA_VERSION=3.6.0] logger.go:42: 07:38:34 | streaming-with-tls/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:38:34 | streaming-with-tls/0-install | >>>> Creating namespace kuttl-test-obliging-maggot logger.go:42: 07:38:34 | streaming-with-tls/0-install | kubectl create namespace kuttl-test-obliging-maggot 2>&1 | grep -v "already exists" || true logger.go:42: 07:38:34 | streaming-with-tls/0-install | >>>> Skipping kafka-operator deployment, assuming it has been installed via OperatorHub logger.go:42: 07:38:34 | streaming-with-tls/0-install | >>>> Creating namespace kuttl-test-obliging-maggot logger.go:42: 07:38:34 | streaming-with-tls/0-install | mkdir -p tests/_build/ logger.go:42: 07:38:34 | streaming-with-tls/0-install | kubectl create namespace kuttl-test-obliging-maggot 2>&1 | grep -v "already exists" || true logger.go:42: 07:38:34 | streaming-with-tls/0-install | curl --fail --location "https://raw.githubusercontent.com/strimzi/strimzi-kafka-operator/3.6.0/examples/kafka/kafka-persistent-single.yaml" --output tests/_build/kafka-example.yaml --create-dirs logger.go:42: 07:38:34 | streaming-with-tls/0-install | % Total % Received % Xferd Average Speed Time Time Time Current logger.go:42: 07:38:34 | streaming-with-tls/0-install | Dload Upload Total Spent Left Speed logger.go:42: 07:38:34 | streaming-with-tls/0-install | 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 14 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 logger.go:42: 07:38:34 | streaming-with-tls/0-install | curl: (22) The requested URL returned error: 404 logger.go:42: 07:38:34 | streaming-with-tls/0-install | make[2]: *** [Makefile:252: kafka] Error 22 logger.go:42: 07:38:34 | streaming-with-tls/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' case.go:364: failed in step 0-install case.go:366: exit status 2 logger.go:42: 07:38:34 | streaming-with-tls | streaming-with-tls events from ns kuttl-test-obliging-maggot: logger.go:42: 07:38:34 | streaming-with-tls | Deleting namespace: kuttl-test-obliging-maggot === CONT kuttl/harness/streaming-simple logger.go:42: 07:38:40 | streaming-simple | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:38:40 | streaming-simple | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:38:40 | streaming-simple | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:38:40 | streaming-simple | Creating namespace: kuttl-test-premium-krill logger.go:42: 07:38:41 | streaming-simple/0-install | starting test step 0-install logger.go:42: 07:38:41 | streaming-simple/0-install | running command: [sh -c cd /tmp/jaeger-tests && make undeploy-kafka KAFKA_NAMESPACE=$NAMESPACE KAFKA_OLM=true] logger.go:42: 07:38:41 | streaming-simple/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:38:41 | streaming-simple/0-install | >>>> Skiping kafka-operator undeploy logger.go:42: 07:38:41 | streaming-simple/0-install | kubectl delete --namespace kuttl-test-premium-krill -f tests/_build/kafka-example.yaml 2>&1 || true logger.go:42: 07:38:41 | streaming-simple/0-install | error: the path "tests/_build/kafka-example.yaml" does not exist logger.go:42: 07:38:41 | streaming-simple/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:38:41 | streaming-simple/0-install | running command: [sh -c cd /tmp/jaeger-tests && make kafka KAFKA_NAMESPACE=$NAMESPACE KAFKA_OLM=true KAFKA_VERSION=3.6.0] logger.go:42: 07:38:41 | streaming-simple/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:38:41 | streaming-simple/0-install | >>>> Creating namespace kuttl-test-premium-krill logger.go:42: 07:38:41 | streaming-simple/0-install | kubectl create namespace kuttl-test-premium-krill 2>&1 | grep -v "already exists" || true logger.go:42: 07:38:41 | streaming-simple/0-install | >>>> Skipping kafka-operator deployment, assuming it has been installed via OperatorHub logger.go:42: 07:38:41 | streaming-simple/0-install | >>>> Creating namespace kuttl-test-premium-krill logger.go:42: 07:38:41 | streaming-simple/0-install | mkdir -p tests/_build/ logger.go:42: 07:38:41 | streaming-simple/0-install | kubectl create namespace kuttl-test-premium-krill 2>&1 | grep -v "already exists" || true logger.go:42: 07:38:41 | streaming-simple/0-install | curl --fail --location "https://raw.githubusercontent.com/strimzi/strimzi-kafka-operator/3.6.0/examples/kafka/kafka-persistent-single.yaml" --output tests/_build/kafka-example.yaml --create-dirs logger.go:42: 07:38:41 | streaming-simple/0-install | % Total % Received % Xferd Average Speed Time Time Time Current logger.go:42: 07:38:41 | streaming-simple/0-install | Dload Upload Total Spent Left Speed logger.go:42: 07:38:41 | streaming-simple/0-install | 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 14 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 logger.go:42: 07:38:41 | streaming-simple/0-install | curl: (22) The requested URL returned error: 404 logger.go:42: 07:38:41 | streaming-simple/0-install | make[2]: *** [Makefile:252: kafka] Error 22 logger.go:42: 07:38:41 | streaming-simple/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' case.go:364: failed in step 0-install case.go:366: exit status 2 logger.go:42: 07:38:41 | streaming-simple | streaming-simple events from ns kuttl-test-premium-krill: logger.go:42: 07:38:41 | streaming-simple | Deleting namespace: kuttl-test-premium-krill === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (214.38s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.16s) --- PASS: kuttl/harness/streaming-with-autoprovisioning-autoscale (194.91s) --- FAIL: kuttl/harness/streaming-with-tls (6.98s) --- FAIL: kuttl/harness/streaming-simple (6.27s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name streaming --report --output /logs/artifacts/streaming.xml ./artifacts/kuttl-report.xml time="2024-01-22T07:38:47Z" level=debug msg="Setting a new name for the test suites" time="2024-01-22T07:38:47Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-01-22T07:38:47Z" level=debug msg="normalizing test case names" time="2024-01-22T07:38:47Z" level=debug msg="streaming/artifacts -> streaming_artifacts" time="2024-01-22T07:38:47Z" level=debug msg="streaming/streaming-with-autoprovisioning-autoscale -> streaming_streaming_with_autoprovisioning_autoscale" time="2024-01-22T07:38:47Z" level=debug msg="streaming/streaming-with-tls -> streaming_streaming_with_tls" time="2024-01-22T07:38:47Z" level=debug msg="streaming/streaming-simple -> streaming_streaming_simple" +-----------------------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------------------+--------+ | streaming_artifacts | passed | | streaming_streaming_with_autoprovisioning_autoscale | passed | | streaming_streaming_with_tls | failed | | streaming_streaming_simple | failed | +-----------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh ui false true + '[' 3 -ne 3 ']' + test_suite_name=ui + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/ui.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-ui make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true ./tests/e2e/ui/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-18-050837 True False 59m Cluster version is 4.15.0-0.nightly-2024-01-18-050837' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-18-050837 True False 59m Cluster version is 4.15.0-0.nightly-2024-01-18-050837' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/ui/render.sh ++ export SUITE_DIR=./tests/e2e/ui ++ SUITE_DIR=./tests/e2e/ui ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/ui ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test allinone + '[' 1 -ne 1 ']' + test_name=allinone + echo =========================================================================== =========================================================================== + info 'Rendering files for test allinone' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test allinone\e[0m' Rendering files for test allinone + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build + '[' _build '!=' _build ']' + mkdir -p allinone + cd allinone + export GET_URL_COMMAND + export URL + export JAEGER_NAME=all-in-one-ui + JAEGER_NAME=all-in-one-ui + '[' true = true ']' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./01-curl.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./04-test-ui-config.yaml + start_test production + '[' 1 -ne 1 ']' + test_name=production + echo =========================================================================== =========================================================================== + info 'Rendering files for test production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test production\e[0m' Rendering files for test production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build/allinone + '[' allinone '!=' _build ']' + cd .. + mkdir -p production + cd production + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + [[ true = true ]] + [[ true = true ]] + render_install_jaeger production-ui production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + '[' true = true ']' + INSECURE=true + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-forbbiden-access.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-curl.yaml + INSECURE=true + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./05-check-disabled-security.yaml + ASSERT_PRESENT=false + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./06-check-NO-gaID.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./08-check-gaID.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running ui E2E tests' Running ui E2E tests + cd tests/e2e/ui/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2576283944 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 3 tests === RUN kuttl/harness === RUN kuttl/harness/allinone === PAUSE kuttl/harness/allinone === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/production === PAUSE kuttl/harness/production === CONT kuttl/harness/allinone logger.go:42: 07:38:54 | allinone | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:38:54 | allinone | Creating namespace: kuttl-test-wanted-haddock logger.go:42: 07:38:54 | allinone/0-install | starting test step 0-install logger.go:42: 07:38:54 | allinone/0-install | Jaeger:kuttl-test-wanted-haddock/all-in-one-ui created logger.go:42: 07:38:58 | allinone/0-install | test step completed 0-install logger.go:42: 07:38:58 | allinone/1-curl | starting test step 1-curl logger.go:42: 07:38:58 | allinone/1-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 07:38:58 | allinone/1-curl | Checking the Ingress host value was populated logger.go:42: 07:38:58 | allinone/1-curl | Try number 0 logger.go:42: 07:38:58 | allinone/1-curl | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 07:38:58 | allinone/1-curl | template was: logger.go:42: 07:38:58 | allinone/1-curl | {.items[0].status.ingress[0].host} logger.go:42: 07:38:58 | allinone/1-curl | object given to jsonpath engine was: logger.go:42: 07:38:58 | allinone/1-curl | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 07:38:58 | allinone/1-curl | logger.go:42: 07:38:58 | allinone/1-curl | logger.go:42: 07:39:08 | allinone/1-curl | Try number 1 logger.go:42: 07:39:08 | allinone/1-curl | Hostname is all-in-one-ui-kuttl-test-wanted-haddock.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:39:08 | allinone/1-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE all-in-one-ui] logger.go:42: 07:39:08 | allinone/1-curl | Checking an expected HTTP response logger.go:42: 07:39:08 | allinone/1-curl | Running in OpenShift logger.go:42: 07:39:08 | allinone/1-curl | User not provided. Getting the token... logger.go:42: 07:39:09 | allinone/1-curl | Warning: resource jaegers/all-in-one-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:39:16 | allinone/1-curl | Try number 1/30 the https://all-in-one-ui-kuttl-test-wanted-haddock.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:39:16 | allinone/1-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:39:16 | allinone/1-curl | Try number 2/30 the https://all-in-one-ui-kuttl-test-wanted-haddock.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:39:16 | allinone/1-curl | curl response asserted properly logger.go:42: 07:39:16 | allinone/1-curl | test step completed 1-curl logger.go:42: 07:39:16 | allinone/2-delete | starting test step 2-delete logger.go:42: 07:39:16 | allinone/2-delete | Jaeger:kuttl-test-wanted-haddock/all-in-one-ui created logger.go:42: 07:39:16 | allinone/2-delete | test step completed 2-delete logger.go:42: 07:39:16 | allinone/3-install | starting test step 3-install logger.go:42: 07:39:16 | allinone/3-install | Jaeger:kuttl-test-wanted-haddock/all-in-one-ui updated logger.go:42: 07:39:16 | allinone/3-install | test step completed 3-install logger.go:42: 07:39:16 | allinone/4-test-ui-config | starting test step 4-test-ui-config logger.go:42: 07:39:16 | allinone/4-test-ui-config | running command: [./ensure-ingress-host.sh] logger.go:42: 07:39:16 | allinone/4-test-ui-config | Checking the Ingress host value was populated logger.go:42: 07:39:16 | allinone/4-test-ui-config | Try number 0 logger.go:42: 07:39:16 | allinone/4-test-ui-config | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 07:39:16 | allinone/4-test-ui-config | template was: logger.go:42: 07:39:16 | allinone/4-test-ui-config | {.items[0].status.ingress[0].host} logger.go:42: 07:39:16 | allinone/4-test-ui-config | object given to jsonpath engine was: logger.go:42: 07:39:16 | allinone/4-test-ui-config | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 07:39:16 | allinone/4-test-ui-config | logger.go:42: 07:39:16 | allinone/4-test-ui-config | logger.go:42: 07:39:26 | allinone/4-test-ui-config | Try number 1 logger.go:42: 07:39:26 | allinone/4-test-ui-config | Hostname is all-in-one-ui-kuttl-test-wanted-haddock.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:39:26 | allinone/4-test-ui-config | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:39:26 | allinone/4-test-ui-config | time="2024-01-22T07:39:26Z" level=info msg="Querying https://all-in-one-ui-kuttl-test-wanted-haddock.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:39:26 | allinone/4-test-ui-config | time="2024-01-22T07:39:26Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:39:26 | allinone/4-test-ui-config | time="2024-01-22T07:39:26Z" level=info msg="Polling to https://all-in-one-ui-kuttl-test-wanted-haddock.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:39:26 | allinone/4-test-ui-config | time="2024-01-22T07:39:26Z" level=info msg="Doing request number 0" logger.go:42: 07:39:26 | allinone/4-test-ui-config | time="2024-01-22T07:39:26Z" level=info msg="Content found and asserted!" logger.go:42: 07:39:26 | allinone/4-test-ui-config | time="2024-01-22T07:39:26Z" level=info msg="Success!" logger.go:42: 07:39:26 | allinone/4-test-ui-config | test step completed 4-test-ui-config logger.go:42: 07:39:26 | allinone | allinone events from ns kuttl-test-wanted-haddock: logger.go:42: 07:39:26 | allinone | 2024-01-22 07:38:57 +0000 UTC Normal Pod all-in-one-ui-759d5694dd-bbj2q Binding Scheduled Successfully assigned kuttl-test-wanted-haddock/all-in-one-ui-759d5694dd-bbj2q to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:39:26 | allinone | 2024-01-22 07:38:57 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-759d5694dd SuccessfulCreate Created pod: all-in-one-ui-759d5694dd-bbj2q replicaset-controller logger.go:42: 07:39:26 | allinone | 2024-01-22 07:38:57 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-759d5694dd to 1 deployment-controller logger.go:42: 07:39:26 | allinone | 2024-01-22 07:38:58 +0000 UTC Normal Pod all-in-one-ui-759d5694dd-bbj2q AddedInterface Add eth0 [10.131.0.77/23] from ovn-kubernetes logger.go:42: 07:39:26 | allinone | 2024-01-22 07:38:58 +0000 UTC Normal Pod all-in-one-ui-759d5694dd-bbj2q.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:38:58 +0000 UTC Normal Pod all-in-one-ui-759d5694dd-bbj2q.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:38:58 +0000 UTC Normal Pod all-in-one-ui-759d5694dd-bbj2q.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:38:58 +0000 UTC Normal Pod all-in-one-ui-759d5694dd-bbj2q.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:38:58 +0000 UTC Normal Pod all-in-one-ui-759d5694dd-bbj2q.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:38:58 +0000 UTC Normal Pod all-in-one-ui-759d5694dd-bbj2q.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:10 +0000 UTC Normal Pod all-in-one-ui-759d5694dd-bbj2q.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:10 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-759d5694dd SuccessfulDelete Deleted pod: all-in-one-ui-759d5694dd-bbj2q replicaset-controller logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:10 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-759d5694dd to 0 from 1 deployment-controller logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:11 +0000 UTC Normal Pod all-in-one-ui-759d5694dd-bbj2q.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:12 +0000 UTC Normal Pod all-in-one-ui-58555845f7-sqgcm Binding Scheduled Successfully assigned kuttl-test-wanted-haddock/all-in-one-ui-58555845f7-sqgcm to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:12 +0000 UTC Normal Pod all-in-one-ui-58555845f7-sqgcm AddedInterface Add eth0 [10.131.0.78/23] from ovn-kubernetes logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:12 +0000 UTC Normal Pod all-in-one-ui-58555845f7-sqgcm.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:12 +0000 UTC Normal Pod all-in-one-ui-58555845f7-sqgcm.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:12 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-58555845f7 SuccessfulCreate Created pod: all-in-one-ui-58555845f7-sqgcm replicaset-controller logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:12 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-58555845f7 to 1 deployment-controller logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:13 +0000 UTC Normal Pod all-in-one-ui-58555845f7-sqgcm.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:13 +0000 UTC Normal Pod all-in-one-ui-58555845f7-sqgcm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:13 +0000 UTC Normal Pod all-in-one-ui-58555845f7-sqgcm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:13 +0000 UTC Normal Pod all-in-one-ui-58555845f7-sqgcm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:16 +0000 UTC Normal Pod all-in-one-ui-58555845f7-sqgcm.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:16 +0000 UTC Normal Pod all-in-one-ui-58555845f7-sqgcm.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:19 +0000 UTC Normal Pod all-in-one-ui-7576b64ff6-xwkg7 Binding Scheduled Successfully assigned kuttl-test-wanted-haddock/all-in-one-ui-7576b64ff6-xwkg7 to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:19 +0000 UTC Warning Pod all-in-one-ui-7576b64ff6-xwkg7 FailedMount MountVolume.SetUp failed for volume "all-in-one-ui-collector-tls-config-volume" : secret "all-in-one-ui-collector-headless-tls" not found kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:19 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-7576b64ff6 SuccessfulCreate Created pod: all-in-one-ui-7576b64ff6-xwkg7 replicaset-controller logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:19 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-7576b64ff6 to 1 deployment-controller logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:20 +0000 UTC Normal Pod all-in-one-ui-7576b64ff6-xwkg7 AddedInterface Add eth0 [10.131.0.79/23] from ovn-kubernetes logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:20 +0000 UTC Normal Pod all-in-one-ui-7576b64ff6-xwkg7.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:a21c5ae4339682904e8075a391a6cee11ddf66142ba0b3e9150c89f1c77ff373" already present on machine kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:21 +0000 UTC Normal Pod all-in-one-ui-7576b64ff6-xwkg7.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:39:26 | allinone | 2024-01-22 07:39:21 +0000 UTC Normal Pod all-in-one-ui-7576b64ff6-xwkg7.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:39:26 | allinone | Deleting namespace: kuttl-test-wanted-haddock === CONT kuttl/harness/production logger.go:42: 07:39:33 | production | Ignoring add-tracking-id.yaml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:39:33 | production | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:39:33 | production | Creating namespace: kuttl-test-massive-muskrat logger.go:42: 07:39:33 | production/1-install | starting test step 1-install logger.go:42: 07:39:33 | production/1-install | Jaeger:kuttl-test-massive-muskrat/production-ui created logger.go:42: 07:40:10 | production/1-install | test step completed 1-install logger.go:42: 07:40:10 | production/2-check-forbbiden-access | starting test step 2-check-forbbiden-access logger.go:42: 07:40:10 | production/2-check-forbbiden-access | running command: [./ensure-ingress-host.sh] logger.go:42: 07:40:10 | production/2-check-forbbiden-access | Checking the Ingress host value was populated logger.go:42: 07:40:10 | production/2-check-forbbiden-access | Try number 0 logger.go:42: 07:40:10 | production/2-check-forbbiden-access | Hostname is production-ui-kuttl-test-massive-muskrat.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:40:10 | production/2-check-forbbiden-access | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE production-ui] logger.go:42: 07:40:11 | production/2-check-forbbiden-access | Checking an expected HTTP response logger.go:42: 07:40:11 | production/2-check-forbbiden-access | Running in OpenShift logger.go:42: 07:40:11 | production/2-check-forbbiden-access | Not using any secret logger.go:42: 07:40:11 | production/2-check-forbbiden-access | Try number 1/30 the https://production-ui-kuttl-test-massive-muskrat.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:40:11 | production/2-check-forbbiden-access | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:40:11 | production/2-check-forbbiden-access | Try number 2/30 the https://production-ui-kuttl-test-massive-muskrat.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:40:11 | production/2-check-forbbiden-access | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 07:40:21 | production/2-check-forbbiden-access | Try number 3/30 the https://production-ui-kuttl-test-massive-muskrat.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:40:21 | production/2-check-forbbiden-access | curl response asserted properly logger.go:42: 07:40:21 | production/2-check-forbbiden-access | test step completed 2-check-forbbiden-access logger.go:42: 07:40:21 | production/3-curl | starting test step 3-curl logger.go:42: 07:40:21 | production/3-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 07:40:21 | production/3-curl | Checking the Ingress host value was populated logger.go:42: 07:40:21 | production/3-curl | Try number 0 logger.go:42: 07:40:21 | production/3-curl | Hostname is production-ui-kuttl-test-massive-muskrat.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:40:21 | production/3-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 07:40:21 | production/3-curl | Checking an expected HTTP response logger.go:42: 07:40:21 | production/3-curl | Running in OpenShift logger.go:42: 07:40:21 | production/3-curl | User not provided. Getting the token... logger.go:42: 07:40:22 | production/3-curl | Warning: resource jaegers/production-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:40:28 | production/3-curl | Try number 1/30 the https://production-ui-kuttl-test-massive-muskrat.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:40:28 | production/3-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:40:28 | production/3-curl | Try number 2/30 the https://production-ui-kuttl-test-massive-muskrat.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:40:28 | production/3-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 07:40:38 | production/3-curl | Try number 3/30 the https://production-ui-kuttl-test-massive-muskrat.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:40:39 | production/3-curl | curl response asserted properly logger.go:42: 07:40:39 | production/3-curl | test step completed 3-curl logger.go:42: 07:40:39 | production/4-install | starting test step 4-install logger.go:42: 07:40:39 | production/4-install | Jaeger:kuttl-test-massive-muskrat/production-ui updated logger.go:42: 07:40:39 | production/4-install | test step completed 4-install logger.go:42: 07:40:39 | production/5-check-disabled-security | starting test step 5-check-disabled-security logger.go:42: 07:40:39 | production/5-check-disabled-security | running command: [./ensure-ingress-host.sh] logger.go:42: 07:40:39 | production/5-check-disabled-security | Checking the Ingress host value was populated logger.go:42: 07:40:39 | production/5-check-disabled-security | Try number 0 logger.go:42: 07:40:39 | production/5-check-disabled-security | Hostname is production-ui-kuttl-test-massive-muskrat.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:40:39 | production/5-check-disabled-security | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 07:40:39 | production/5-check-disabled-security | Checking an expected HTTP response logger.go:42: 07:40:39 | production/5-check-disabled-security | Running in OpenShift logger.go:42: 07:40:39 | production/5-check-disabled-security | Not using any secret logger.go:42: 07:40:39 | production/5-check-disabled-security | Try number 1/30 the https://production-ui-kuttl-test-massive-muskrat.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:40:39 | production/5-check-disabled-security | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:40:39 | production/5-check-disabled-security | Try number 2/30 the https://production-ui-kuttl-test-massive-muskrat.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:40:39 | production/5-check-disabled-security | HTTP response is 403. 200 expected. Waiting 10 s logger.go:42: 07:40:49 | production/5-check-disabled-security | Try number 3/30 the https://production-ui-kuttl-test-massive-muskrat.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search logger.go:42: 07:40:49 | production/5-check-disabled-security | curl response asserted properly logger.go:42: 07:40:49 | production/5-check-disabled-security | test step completed 5-check-disabled-security logger.go:42: 07:40:49 | production/6-check-NO-gaID | starting test step 6-check-NO-gaID logger.go:42: 07:40:49 | production/6-check-NO-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 07:40:49 | production/6-check-NO-gaID | Checking the Ingress host value was populated logger.go:42: 07:40:49 | production/6-check-NO-gaID | Try number 0 logger.go:42: 07:40:49 | production/6-check-NO-gaID | Hostname is production-ui-kuttl-test-massive-muskrat.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:40:49 | production/6-check-NO-gaID | running command: [sh -c ASSERT_PRESENT=false EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:40:49 | production/6-check-NO-gaID | time="2024-01-22T07:40:49Z" level=info msg="Querying https://production-ui-kuttl-test-massive-muskrat.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:40:49 | production/6-check-NO-gaID | time="2024-01-22T07:40:49Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:40:49 | production/6-check-NO-gaID | time="2024-01-22T07:40:49Z" level=info msg="Polling to https://production-ui-kuttl-test-massive-muskrat.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:40:49 | production/6-check-NO-gaID | time="2024-01-22T07:40:49Z" level=info msg="Doing request number 0" logger.go:42: 07:40:49 | production/6-check-NO-gaID | time="2024-01-22T07:40:49Z" level=info msg="Content not found and asserted it was not found!" logger.go:42: 07:40:49 | production/6-check-NO-gaID | time="2024-01-22T07:40:49Z" level=info msg="Success!" logger.go:42: 07:40:49 | production/6-check-NO-gaID | test step completed 6-check-NO-gaID logger.go:42: 07:40:49 | production/7-add-tracking-id | starting test step 7-add-tracking-id logger.go:42: 07:40:49 | production/7-add-tracking-id | running command: [sh -c kubectl apply -f add-tracking-id.yaml -n $NAMESPACE] logger.go:42: 07:40:50 | production/7-add-tracking-id | jaeger.jaegertracing.io/production-ui configured logger.go:42: 07:40:50 | production/7-add-tracking-id | test step completed 7-add-tracking-id logger.go:42: 07:40:50 | production/8-check-gaID | starting test step 8-check-gaID logger.go:42: 07:40:50 | production/8-check-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 07:40:50 | production/8-check-gaID | Checking the Ingress host value was populated logger.go:42: 07:40:50 | production/8-check-gaID | Try number 0 logger.go:42: 07:40:50 | production/8-check-gaID | Hostname is production-ui-kuttl-test-massive-muskrat.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com logger.go:42: 07:40:50 | production/8-check-gaID | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:40:50 | production/8-check-gaID | time="2024-01-22T07:40:50Z" level=info msg="Querying https://production-ui-kuttl-test-massive-muskrat.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:40:50 | production/8-check-gaID | time="2024-01-22T07:40:50Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:40:50 | production/8-check-gaID | time="2024-01-22T07:40:50Z" level=info msg="Polling to https://production-ui-kuttl-test-massive-muskrat.apps.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:40:50 | production/8-check-gaID | time="2024-01-22T07:40:50Z" level=info msg="Doing request number 0" logger.go:42: 07:40:50 | production/8-check-gaID | time="2024-01-22T07:40:50Z" level=warning msg="Found: false . Assert: true" logger.go:42: 07:40:50 | production/8-check-gaID | time="2024-01-22T07:40:50Z" level=warning msg="The condition of the test function was not accomplished" logger.go:42: 07:40:58 | production/8-check-gaID | time="2024-01-22T07:40:58Z" level=info msg="Doing request number 1" logger.go:42: 07:40:58 | production/8-check-gaID | time="2024-01-22T07:40:58Z" level=info msg="Content found and asserted!" logger.go:42: 07:40:58 | production/8-check-gaID | time="2024-01-22T07:40:58Z" level=info msg="Success!" logger.go:42: 07:40:58 | production/8-check-gaID | test step completed 8-check-gaID logger.go:42: 07:40:58 | production | production events from ns kuttl-test-massive-muskrat: logger.go:42: 07:40:58 | production | 2024-01-22 07:39:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmassivemuskratproductionui-1-6b9f4hp Binding Scheduled Successfully assigned kuttl-test-massive-muskrat/elasticsearch-cdm-kuttltestmassivemuskratproductionui-1-6b9f4hp to ip-10-0-102-115.ec2.internal default-scheduler logger.go:42: 07:40:58 | production | 2024-01-22 07:39:39 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmassivemuskratproductionui-1-6b9f4hp FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:39:39 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestmassivemuskratproductionui-1-6bf6f8dbb7 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestmassivemuskratproductionui-1-6b9f4hp replicaset-controller logger.go:42: 07:40:58 | production | 2024-01-22 07:39:39 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestmassivemuskratproductionui-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestmassivemuskratproductionui-1-6bf6f8dbb7 to 1 deployment-controller logger.go:42: 07:40:58 | production | 2024-01-22 07:39:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmassivemuskratproductionui-1-6b9f4hp AddedInterface Add eth0 [10.131.0.80/23] from ovn-kubernetes logger.go:42: 07:40:58 | production | 2024-01-22 07:39:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmassivemuskratproductionui-1-6b9f4hp.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:3ad088cbfcfb006d87762e96a70ff989e75d49872954347f7efe34d9529c1b7a" already present on machine kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:39:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmassivemuskratproductionui-1-6b9f4hp.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:39:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmassivemuskratproductionui-1-6b9f4hp.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:39:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmassivemuskratproductionui-1-6b9f4hp.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:911ee3f26ae71eab71bf2a1455d1c2ea871cddc78a51189e3ad2f18f63bd3d91" already present on machine kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:39:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmassivemuskratproductionui-1-6b9f4hp.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:39:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmassivemuskratproductionui-1-6b9f4hp.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:39:50 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmassivemuskratproductionui-1-6b9f4hp.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:39:55 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmassivemuskratproductionui-1-6b9f4hp.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:06 +0000 UTC Normal ReplicaSet.apps production-ui-collector-56ffcb6849 SuccessfulCreate Created pod: production-ui-collector-56ffcb6849-nqk42 replicaset-controller logger.go:42: 07:40:58 | production | 2024-01-22 07:40:06 +0000 UTC Normal Deployment.apps production-ui-collector ScalingReplicaSet Scaled up replica set production-ui-collector-56ffcb6849 to 1 deployment-controller logger.go:42: 07:40:58 | production | 2024-01-22 07:40:06 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-6f4556876b to 1 deployment-controller logger.go:42: 07:40:58 | production | 2024-01-22 07:40:07 +0000 UTC Normal Pod production-ui-collector-56ffcb6849-nqk42 Binding Scheduled Successfully assigned kuttl-test-massive-muskrat/production-ui-collector-56ffcb6849-nqk42 to ip-10-0-15-245.ec2.internal default-scheduler logger.go:42: 07:40:58 | production | 2024-01-22 07:40:07 +0000 UTC Normal Pod production-ui-collector-56ffcb6849-nqk42 AddedInterface Add eth0 [10.128.2.102/23] from ovn-kubernetes logger.go:42: 07:40:58 | production | 2024-01-22 07:40:07 +0000 UTC Normal Pod production-ui-collector-56ffcb6849-nqk42.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:9ec63242b45d2dd0e95ce79b5f0e04736cda1363b76c9c0d7453e92a26442fd9" already present on machine kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:07 +0000 UTC Normal Pod production-ui-collector-56ffcb6849-nqk42.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:07 +0000 UTC Normal Pod production-ui-collector-56ffcb6849-nqk42.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:07 +0000 UTC Normal Pod production-ui-query-6f4556876b-7vhg9 Binding Scheduled Successfully assigned kuttl-test-massive-muskrat/production-ui-query-6f4556876b-7vhg9 to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:40:58 | production | 2024-01-22 07:40:07 +0000 UTC Normal Pod production-ui-query-6f4556876b-7vhg9 AddedInterface Add eth0 [10.129.2.59/23] from ovn-kubernetes logger.go:42: 07:40:58 | production | 2024-01-22 07:40:07 +0000 UTC Normal Pod production-ui-query-6f4556876b-7vhg9.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:07 +0000 UTC Normal Pod production-ui-query-6f4556876b-7vhg9.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:07 +0000 UTC Normal Pod production-ui-query-6f4556876b-7vhg9.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:07 +0000 UTC Normal Pod production-ui-query-6f4556876b-7vhg9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:07 +0000 UTC Normal Pod production-ui-query-6f4556876b-7vhg9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:07 +0000 UTC Normal Pod production-ui-query-6f4556876b-7vhg9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:07 +0000 UTC Normal Pod production-ui-query-6f4556876b-7vhg9.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:07 +0000 UTC Normal Pod production-ui-query-6f4556876b-7vhg9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:07 +0000 UTC Normal Pod production-ui-query-6f4556876b-7vhg9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:07 +0000 UTC Normal ReplicaSet.apps production-ui-query-6f4556876b SuccessfulCreate Created pod: production-ui-query-6f4556876b-7vhg9 replicaset-controller logger.go:42: 07:40:58 | production | 2024-01-22 07:40:22 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:40:58 | production | 2024-01-22 07:40:22 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:40:58 | production | 2024-01-22 07:40:22 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:40:58 | production | 2024-01-22 07:40:23 +0000 UTC Normal Pod production-ui-query-6f4556876b-7vhg9.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:23 +0000 UTC Normal Pod production-ui-query-6f4556876b-7vhg9.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:23 +0000 UTC Normal ReplicaSet.apps production-ui-query-6f4556876b SuccessfulDelete Deleted pod: production-ui-query-6f4556876b-7vhg9 replicaset-controller logger.go:42: 07:40:58 | production | 2024-01-22 07:40:23 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-6f4556876b to 0 from 1 deployment-controller logger.go:42: 07:40:58 | production | 2024-01-22 07:40:24 +0000 UTC Normal Pod production-ui-query-6f4556876b-7vhg9.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:25 +0000 UTC Normal Pod production-ui-query-84c4bb9cd4-rblg5 Binding Scheduled Successfully assigned kuttl-test-massive-muskrat/production-ui-query-84c4bb9cd4-rblg5 to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:40:58 | production | 2024-01-22 07:40:25 +0000 UTC Normal Pod production-ui-query-84c4bb9cd4-rblg5 AddedInterface Add eth0 [10.129.2.60/23] from ovn-kubernetes logger.go:42: 07:40:58 | production | 2024-01-22 07:40:25 +0000 UTC Normal Pod production-ui-query-84c4bb9cd4-rblg5.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:25 +0000 UTC Normal Pod production-ui-query-84c4bb9cd4-rblg5.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:25 +0000 UTC Normal Pod production-ui-query-84c4bb9cd4-rblg5.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:25 +0000 UTC Normal Pod production-ui-query-84c4bb9cd4-rblg5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:219bf2d14157acd90298df58bfe77c2e3ed51ce0c743c2e51b3ed54b73dafc14" already present on machine kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:25 +0000 UTC Normal Pod production-ui-query-84c4bb9cd4-rblg5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:25 +0000 UTC Normal Pod production-ui-query-84c4bb9cd4-rblg5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:25 +0000 UTC Normal Pod production-ui-query-84c4bb9cd4-rblg5.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:25 +0000 UTC Normal ReplicaSet.apps production-ui-query-84c4bb9cd4 SuccessfulCreate Created pod: production-ui-query-84c4bb9cd4-rblg5 replicaset-controller logger.go:42: 07:40:58 | production | 2024-01-22 07:40:25 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-84c4bb9cd4 to 1 deployment-controller logger.go:42: 07:40:58 | production | 2024-01-22 07:40:26 +0000 UTC Normal Pod production-ui-query-84c4bb9cd4-rblg5.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:26 +0000 UTC Normal Pod production-ui-query-84c4bb9cd4-rblg5.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:40 +0000 UTC Normal Pod production-ui-query-84c4bb9cd4-rblg5.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:40 +0000 UTC Normal Pod production-ui-query-84c4bb9cd4-rblg5.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:40 +0000 UTC Normal Pod production-ui-query-84c4bb9cd4-rblg5.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:40 +0000 UTC Normal ReplicaSet.apps production-ui-query-84c4bb9cd4 SuccessfulDelete Deleted pod: production-ui-query-84c4bb9cd4-rblg5 replicaset-controller logger.go:42: 07:40:58 | production | 2024-01-22 07:40:40 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-84c4bb9cd4 to 0 from 1 deployment-controller logger.go:42: 07:40:58 | production | 2024-01-22 07:40:41 +0000 UTC Normal Pod production-ui-query-778fb765b5-j87hd Binding Scheduled Successfully assigned kuttl-test-massive-muskrat/production-ui-query-778fb765b5-j87hd to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:40:58 | production | 2024-01-22 07:40:41 +0000 UTC Normal ReplicaSet.apps production-ui-query-778fb765b5 SuccessfulCreate Created pod: production-ui-query-778fb765b5-j87hd replicaset-controller logger.go:42: 07:40:58 | production | 2024-01-22 07:40:41 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-778fb765b5 to 1 deployment-controller logger.go:42: 07:40:58 | production | 2024-01-22 07:40:42 +0000 UTC Warning Pod production-ui-query-778fb765b5-j87hd FailedMount MountVolume.SetUp failed for volume "production-ui-service-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:42 +0000 UTC Warning Pod production-ui-query-778fb765b5-j87hd FailedMount MountVolume.SetUp failed for volume "production-ui-trusted-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:43 +0000 UTC Normal Pod production-ui-query-778fb765b5-j87hd AddedInterface Add eth0 [10.129.2.61/23] from ovn-kubernetes logger.go:42: 07:40:58 | production | 2024-01-22 07:40:43 +0000 UTC Normal Pod production-ui-query-778fb765b5-j87hd.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:43 +0000 UTC Normal Pod production-ui-query-778fb765b5-j87hd.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:43 +0000 UTC Normal Pod production-ui-query-778fb765b5-j87hd.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:43 +0000 UTC Normal Pod production-ui-query-778fb765b5-j87hd.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:43 +0000 UTC Normal Pod production-ui-query-778fb765b5-j87hd.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:43 +0000 UTC Normal Pod production-ui-query-778fb765b5-j87hd.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:51 +0000 UTC Normal Pod production-ui-query-778fb765b5-j87hd.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:51 +0000 UTC Normal Pod production-ui-query-778fb765b5-j87hd.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:51 +0000 UTC Normal ReplicaSet.apps production-ui-query-778fb765b5 SuccessfulDelete Deleted pod: production-ui-query-778fb765b5-j87hd replicaset-controller logger.go:42: 07:40:58 | production | 2024-01-22 07:40:51 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-778fb765b5 to 0 from 1 deployment-controller logger.go:42: 07:40:58 | production | 2024-01-22 07:40:52 +0000 UTC Normal Pod production-ui-query-d98bc66c8-dcb4f Binding Scheduled Successfully assigned kuttl-test-massive-muskrat/production-ui-query-d98bc66c8-dcb4f to ip-10-0-43-114.ec2.internal default-scheduler logger.go:42: 07:40:58 | production | 2024-01-22 07:40:52 +0000 UTC Normal Pod production-ui-query-d98bc66c8-dcb4f AddedInterface Add eth0 [10.129.2.62/23] from ovn-kubernetes logger.go:42: 07:40:58 | production | 2024-01-22 07:40:52 +0000 UTC Normal Pod production-ui-query-d98bc66c8-dcb4f.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:116963b147bccac665fa10f28107468699d5ef632b5c86710a3e900423b404a9" already present on machine kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:52 +0000 UTC Normal Pod production-ui-query-d98bc66c8-dcb4f.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:52 +0000 UTC Normal Pod production-ui-query-d98bc66c8-dcb4f.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:52 +0000 UTC Normal Pod production-ui-query-d98bc66c8-dcb4f.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:4437924ff09250ff83e94a4afa7bcbd46231e10ab55747023fe9c072d1484c79" already present on machine kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:52 +0000 UTC Normal Pod production-ui-query-d98bc66c8-dcb4f.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:52 +0000 UTC Normal Pod production-ui-query-d98bc66c8-dcb4f.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:40:58 | production | 2024-01-22 07:40:52 +0000 UTC Normal ReplicaSet.apps production-ui-query-d98bc66c8 SuccessfulCreate Created pod: production-ui-query-d98bc66c8-dcb4f replicaset-controller logger.go:42: 07:40:58 | production | 2024-01-22 07:40:52 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-d98bc66c8 to 1 deployment-controller logger.go:42: 07:40:58 | production | Deleting namespace: kuttl-test-massive-muskrat === CONT kuttl/harness/artifacts logger.go:42: 07:41:05 | artifacts | Creating namespace: kuttl-test-credible-ostrich logger.go:42: 07:41:05 | artifacts | artifacts events from ns kuttl-test-credible-ostrich: logger.go:42: 07:41:05 | artifacts | Deleting namespace: kuttl-test-credible-ostrich === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (136.95s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/allinone (39.25s) --- PASS: kuttl/harness/production (91.70s) --- PASS: kuttl/harness/artifacts (5.96s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name ui --report --output /logs/artifacts/ui.xml ./artifacts/kuttl-report.xml time="2024-01-22T07:41:11Z" level=debug msg="Setting a new name for the test suites" time="2024-01-22T07:41:11Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-01-22T07:41:11Z" level=debug msg="normalizing test case names" time="2024-01-22T07:41:11Z" level=debug msg="ui/allinone -> ui_allinone" time="2024-01-22T07:41:11Z" level=debug msg="ui/production -> ui_production" time="2024-01-22T07:41:11Z" level=debug msg="ui/artifacts -> ui_artifacts" +---------------+--------+ | NAME | RESULT | +---------------+--------+ | ui_allinone | passed | | ui_production | passed | | ui_artifacts | passed | +---------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh upgrade false true + '[' 3 -ne 3 ']' + test_suite_name=upgrade + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/upgrade.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-upgrade make[2]: Entering directory '/tmp/jaeger-tests' make docker JAEGER_VERSION=1.53.1 IMG="quay.io//jaeger-operator:next" make[3]: Entering directory '/tmp/jaeger-tests' [ ! -z "true" ] || docker build --build-arg=GOPROXY= --build-arg=VERSION="1.53.0" --build-arg=JAEGER_VERSION=1.53.1 --build-arg=TARGETARCH= --build-arg VERSION_DATE=2024-01-22T07:41:11Z --build-arg VERSION_PKG="github.com/jaegertracing/jaeger-operator/pkg/version" -t "quay.io//jaeger-operator:next" . make[3]: Leaving directory '/tmp/jaeger-tests' touch build-e2e-upgrade-image SKIP_ES_EXTERNAL=true IMG=quay.io//jaeger-operator:"1.53.0" JAEGER_OPERATOR_VERSION="1.53.0" JAEGER_VERSION="1.53.0" ./tests/e2e/upgrade/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-18-050837 True False 62m Cluster version is 4.15.0-0.nightly-2024-01-18-050837' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.0-0.nightly-2024-01-18-050837 True False 62m Cluster version is 4.15.0-0.nightly-2024-01-18-050837' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/upgrade/render.sh ++ export SUITE_DIR=./tests/e2e/upgrade ++ SUITE_DIR=./tests/e2e/upgrade ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/upgrade ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + export JAEGER_NAME + '[' true = true ']' + skip_test upgrade 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade + warning 'upgrade: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade: Test not supported in OpenShift\e[0m' WAR: upgrade: Test not supported in OpenShift + '[' true = true ']' + skip_test upgrade-from-latest-release 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade-from-latest-release + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade-from-latest-release + warning 'upgrade-from-latest-release: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade-from-latest-release: Test not supported in OpenShift\e[0m' WAR: upgrade-from-latest-release: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running upgrade E2E tests' Running upgrade E2E tests + cd tests/e2e/upgrade/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2576283944 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-zz0glpxl-5054a.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:41:12 | artifacts | Creating namespace: kuttl-test-mighty-falcon logger.go:42: 07:41:12 | artifacts | artifacts events from ns kuttl-test-mighty-falcon: logger.go:42: 07:41:12 | artifacts | Deleting namespace: kuttl-test-mighty-falcon === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (5.89s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.84s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name upgrade --report --output /logs/artifacts/upgrade.xml ./artifacts/kuttl-report.xml time="2024-01-22T07:41:18Z" level=debug msg="Setting a new name for the test suites" time="2024-01-22T07:41:18Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-01-22T07:41:18Z" level=debug msg="normalizing test case names" time="2024-01-22T07:41:18Z" level=debug msg="upgrade/artifacts -> upgrade_artifacts" +-------------------+--------+ | NAME | RESULT | +-------------------+--------+ | upgrade_artifacts | passed | +-------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests'