% Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 3831 100 3831 0 0 34709 0 --:--:-- --:--:-- --:--:-- 34827 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 9315 100 9315 0 0 100k 0 --:--:-- --:--:-- --:--:-- 101k % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 2607 100 2607 0 0 24070 0 --:--:-- --:--:-- --:--:-- 24138 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 1953 100 1953 0 0 24240 0 --:--:-- --:--:-- --:--:-- 24412 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 350 100 350 0 0 3526 0 --:--:-- --:--:-- --:--:-- 3535 Installing kuttl Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/kubectl-kuttl https://github.com/kudobuilder/kuttl/releases/download/v0.15.0/kubectl-kuttl_0.15.0_linux_x86_64 KUBECONFIG file is: /tmp/kubeconfig-3357233117 for suite in elasticsearch examples generate upgrade sidecar streaming ui miscellaneous; do \ make run-e2e-tests-$suite ; \ done make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh elasticsearch false true + '[' 3 -ne 3 ']' + test_suite_name=elasticsearch + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/elasticsearch.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-elasticsearch make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true \ KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ ./tests/e2e/elasticsearch/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-07-024651 True False 11m Cluster version is 4.16.0-0.nightly-2024-06-07-024651' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-07-024651 True False 11m Cluster version is 4.16.0-0.nightly-2024-06-07-024651' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/elasticsearch/render.sh ++ export SUITE_DIR=./tests/e2e/elasticsearch ++ SUITE_DIR=./tests/e2e/elasticsearch ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/elasticsearch ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + start_test es-from-aio-to-production + '[' 1 -ne 1 ']' + test_name=es-from-aio-to-production + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-from-aio-to-production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-from-aio-to-production\e[0m' Rendering files for test es-from-aio-to-production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-from-aio-to-production + cd es-from-aio-to-production + jaeger_name=my-jaeger + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 03 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=03 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./03-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch.redundancyPolicy="ZeroRedundancy"' ./03-install.yaml + render_smoke_test my-jaeger true 04 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=04 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./04-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test es-increasing-replicas + '[' 1 -ne 1 ']' + test_name=es-increasing-replicas + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-increasing-replicas' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-increasing-replicas\e[0m' Rendering files for test es-increasing-replicas + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-from-aio-to-production + '[' es-from-aio-to-production '!=' _build ']' + cd .. + mkdir -p es-increasing-replicas + cd es-increasing-replicas + jaeger_name=simple-prod + '[' true = true ']' + jaeger_deployment_mode=production_autoprovisioned + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.replicas=2 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.query.replicas=2 ./02-install.yaml + cp ./01-assert.yaml ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.replicas=2 ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .status.readyReplicas=2 ./02-assert.yaml + render_smoke_test simple-prod true 03 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=03 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./03-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + cp ./02-install.yaml ./04-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.elasticsearch.nodeCount=2 ./04-install.yaml + /tmp/jaeger-tests/bin/gomplate -f ./openshift-check-es-nodes.yaml.template -o ./05-check-es-nodes.yaml + '[' true = true ']' + skip_test es-index-cleaner-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-increasing-replicas + '[' es-increasing-replicas '!=' _build ']' + cd .. + rm -rf es-index-cleaner-upstream + warning 'es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_index_cleaner -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-index-cleaner-autoprov + '[' 1 -ne 1 ']' + test_name=es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-index-cleaner-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-index-cleaner-autoprov\e[0m' Rendering files for test es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-index-cleaner-autoprov + cd es-index-cleaner-autoprov + jaeger_name=test-es-index-cleaner-with-prefix + cronjob_name=test-es-index-cleaner-with-prefix-es-index-cleaner + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + cp ../../es-index-cleaner-upstream/04-assert.yaml ../../es-index-cleaner-upstream/README.md . + render_install_jaeger test-es-index-cleaner-with-prefix production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options.es.index-prefix=""' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.enabled=false ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.numberOfDays=0 ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.esIndexCleaner.schedule="*/1 * * * *"' ./01-install.yaml + render_report_spans test-es-index-cleaner-with-prefix true 5 00 true 02 + '[' 6 -ne 6 ']' + jaeger=test-es-index-cleaner-with-prefix + is_secured=true + number_of_spans=5 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + export JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=5 + DAYS=5 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + sed 's~enabled: false~enabled: true~gi' ./01-install.yaml + CRONJOB_NAME=test-es-index-cleaner-with-prefix-es-index-cleaner + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./04-wait-es-index-cleaner.yaml + /tmp/jaeger-tests/bin/gomplate -f ./01-install.yaml -o ./05-install.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 00 06 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=00 + test_step=06 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=test-es-index-cleaner-with-prefix-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=test-es-index-cleaner-with-prefix-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./06-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./06-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.3 ++ version_ge 5.8.3 5.4 +++ echo 5.8.3 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.3 == 5.8.3 + '[' -n '' ']' + skip_test es-index-cleaner-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-index-cleaner-autoprov + '[' es-index-cleaner-autoprov '!=' _build ']' + cd .. + rm -rf es-index-cleaner-managed + warning 'es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + start_test es-multiinstance + '[' 1 -ne 1 ']' + test_name=es-multiinstance + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-multiinstance' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-multiinstance\e[0m' Rendering files for test es-multiinstance + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-multiinstance + cd es-multiinstance + jaeger_name=instance-1 + render_install_jaeger instance-1 production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=instance-1 + JAEGER_NAME=instance-1 + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f ./03-create-second-instance.yaml.template -o 03-create-second-instance.yaml + '[' true = true ']' + skip_test es-rollover-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-rollover-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-multiinstance + '[' es-multiinstance '!=' _build ']' + cd .. + rm -rf es-rollover-upstream + warning 'es-rollover-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_rollover -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-rollover-autoprov + '[' 1 -ne 1 ']' + test_name=es-rollover-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-rollover-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-rollover-autoprov\e[0m' Rendering files for test es-rollover-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-rollover-autoprov + cd es-rollover-autoprov + cp ../../es-rollover-upstream/05-assert.yaml ../../es-rollover-upstream/05-install.yaml ../../es-rollover-upstream/README.md . + jaeger_name=my-jaeger + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_report_spans my-jaeger true 2 00 true 02 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 00 03 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=00 + test_step=03 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./03-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./03-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 01 04 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=01 + test_step=04 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=01 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./04-check-indices.yaml + JOB_NUMBER=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./04-assert.yaml + render_report_spans my-jaeger true 2 02 true 06 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=02 + ensure_reported_spans=true + test_step=06 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=02 + JOB_NUMBER=02 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./06-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./06-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 02 07 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=02 + test_step=07 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=02 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./07-check-indices.yaml + JOB_NUMBER=02 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./07-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' 03 08 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + job_number=03 + test_step=08 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=03 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./08-check-indices.yaml + JOB_NUMBER=03 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./08-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' 04 09 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + job_number=04 + test_step=09 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=04 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./09-check-indices.yaml + JOB_NUMBER=04 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./09-assert.yaml + render_report_spans my-jaeger true 2 03 true 10 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=03 + ensure_reported_spans=true + test_step=10 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=03 + JOB_NUMBER=03 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./10-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./10-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + CRONJOB_NAME=my-jaeger-es-rollover + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./11-wait-rollover.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-000002'\'',' 05 11 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-000002'\'',' + job_number=05 + test_step=11 + escape_command ''\''--name'\'', '\''jaeger-span-000002'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-000002'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-000002'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-000002'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=05 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./11-check-indices.yaml + JOB_NUMBER=05 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./11-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' 06 12 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + job_number=06 + test_step=12 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=06 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./12-check-indices.yaml + JOB_NUMBER=06 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./12-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.3 ++ version_ge 5.8.3 5.4 +++ echo 5.8.3 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.3 == 5.8.3 + '[' -n '' ']' + skip_test es-rollover-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-rollover-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-rollover-autoprov + '[' es-rollover-autoprov '!=' _build ']' + cd .. + rm -rf es-rollover-managed + warning 'es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + skip_test es-spark-dependencies 'This test is not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=es-spark-dependencies + message='This test is not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + rm -rf es-spark-dependencies + warning 'es-spark-dependencies: This test is not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-spark-dependencies: This test is not supported in OpenShift\e[0m' WAR: es-spark-dependencies: This test is not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running elasticsearch E2E tests' Running elasticsearch E2E tests + cd tests/e2e/elasticsearch/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3357233117 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/es-from-aio-to-production === PAUSE kuttl/harness/es-from-aio-to-production === RUN kuttl/harness/es-increasing-replicas === PAUSE kuttl/harness/es-increasing-replicas === RUN kuttl/harness/es-index-cleaner-autoprov === PAUSE kuttl/harness/es-index-cleaner-autoprov === RUN kuttl/harness/es-multiinstance === PAUSE kuttl/harness/es-multiinstance === RUN kuttl/harness/es-rollover-autoprov === PAUSE kuttl/harness/es-rollover-autoprov === RUN kuttl/harness/es-simple-prod === PAUSE kuttl/harness/es-simple-prod === CONT kuttl/harness/artifacts logger.go:42: 06:58:11 | artifacts | Creating namespace: kuttl-test-heroic-osprey logger.go:42: 06:58:11 | artifacts | artifacts events from ns kuttl-test-heroic-osprey: logger.go:42: 06:58:11 | artifacts | Deleting namespace: kuttl-test-heroic-osprey === CONT kuttl/harness/es-multiinstance logger.go:42: 06:58:17 | es-multiinstance | Ignoring 03-create-second-instance.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 06:58:17 | es-multiinstance | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 06:58:17 | es-multiinstance | Creating namespace: kuttl-test-meet-fox logger.go:42: 06:58:17 | es-multiinstance/0-clear-namespace | starting test step 0-clear-namespace logger.go:42: 06:58:17 | es-multiinstance/0-clear-namespace | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --ignore-not-found=true] logger.go:42: 06:58:18 | es-multiinstance/0-clear-namespace | test step completed 0-clear-namespace logger.go:42: 06:58:18 | es-multiinstance/1-install | starting test step 1-install logger.go:42: 06:58:18 | es-multiinstance/1-install | Jaeger:kuttl-test-meet-fox/instance-1 created logger.go:42: 06:59:13 | es-multiinstance/1-install | test step completed 1-install logger.go:42: 06:59:13 | es-multiinstance/2-create-namespace | starting test step 2-create-namespace logger.go:42: 06:59:13 | es-multiinstance/2-create-namespace | running command: [sh -c kubectl create namespace jaeger-e2e-multiinstance-test] logger.go:42: 06:59:13 | es-multiinstance/2-create-namespace | namespace/jaeger-e2e-multiinstance-test created logger.go:42: 06:59:13 | es-multiinstance/2-create-namespace | test step completed 2-create-namespace logger.go:42: 06:59:13 | es-multiinstance/3-create-second-instance | starting test step 3-create-second-instance logger.go:42: 06:59:13 | es-multiinstance/3-create-second-instance | running command: [sh -c kubectl apply -f ./01-install.yaml -n jaeger-e2e-multiinstance-test] logger.go:42: 06:59:15 | es-multiinstance/3-create-second-instance | jaeger.jaegertracing.io/instance-1 created logger.go:42: 06:59:15 | es-multiinstance/3-create-second-instance | running command: [sh -c /tmp/jaeger-tests/bin/kubectl-kuttl assert ./01-assert.yaml -n jaeger-e2e-multiinstance-test --timeout 1000] logger.go:42: 07:00:07 | es-multiinstance/3-create-second-instance | assert is valid logger.go:42: 07:00:07 | es-multiinstance/3-create-second-instance | test step completed 3-create-second-instance logger.go:42: 07:00:07 | es-multiinstance/4-check-secrets | starting test step 4-check-secrets logger.go:42: 07:00:07 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n $NAMESPACE > secret1] logger.go:42: 07:00:07 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n jaeger-e2e-multiinstance-test > secret2] logger.go:42: 07:00:08 | es-multiinstance/4-check-secrets | running command: [sh -c cmp --silent secret1 secret2 || exit 0] logger.go:42: 07:00:08 | es-multiinstance/4-check-secrets | test step completed 4-check-secrets logger.go:42: 07:00:08 | es-multiinstance/5-delete | starting test step 5-delete logger.go:42: 07:00:08 | es-multiinstance/5-delete | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --wait=false] logger.go:42: 07:00:08 | es-multiinstance/5-delete | namespace "jaeger-e2e-multiinstance-test" deleted logger.go:42: 07:00:08 | es-multiinstance/5-delete | test step completed 5-delete logger.go:42: 07:00:08 | es-multiinstance | es-multiinstance events from ns kuttl-test-meet-fox: logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:58:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeetfoxinstance1-1-c8699d49b-vh7bz Binding Scheduled Successfully assigned kuttl-test-meet-fox/elasticsearch-cdm-kuttltestmeetfoxinstance1-1-c8699d49b-vh7bz to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:58:26 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestmeetfoxinstance1-1-c8699d49b SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestmeetfoxinstance1-1-c8699d49b-vh7bz replicaset-controller logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:58:26 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestmeetfoxinstance1-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestmeetfoxinstance1-1-c8699d49b to 1 deployment-controller logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:58:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeetfoxinstance1-1-c8699d49b-vh7bz AddedInterface Add eth0 [10.131.0.21/23] from ovn-kubernetes multus logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:58:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeetfoxinstance1-1-c8699d49b-vh7bz.spec.containers{elasticsearch} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:58:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeetfoxinstance1-1-c8699d49b-vh7bz.spec.containers{elasticsearch} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" in 7.701s (7.701s including waiting) kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:58:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeetfoxinstance1-1-c8699d49b-vh7bz.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:58:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeetfoxinstance1-1-c8699d49b-vh7bz.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:58:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeetfoxinstance1-1-c8699d49b-vh7bz.spec.containers{proxy} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:58:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeetfoxinstance1-1-c8699d49b-vh7bz.spec.containers{proxy} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" in 3.162s (3.162s including waiting) kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:58:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeetfoxinstance1-1-c8699d49b-vh7bz.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:58:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeetfoxinstance1-1-c8699d49b-vh7bz.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:58:47 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmeetfoxinstance1-1-c8699d49b-vh7bz.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:58:52 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmeetfoxinstance1-1-c8699d49b-vh7bz.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:03 +0000 UTC Normal Pod instance-1-collector-85884cd4-nq4bz Binding Scheduled Successfully assigned kuttl-test-meet-fox/instance-1-collector-85884cd4-nq4bz to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:03 +0000 UTC Normal ReplicaSet.apps instance-1-collector-85884cd4 SuccessfulCreate Created pod: instance-1-collector-85884cd4-nq4bz replicaset-controller logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:03 +0000 UTC Normal Deployment.apps instance-1-collector ScalingReplicaSet Scaled up replica set instance-1-collector-85884cd4 to 1 deployment-controller logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:03 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-8db556c94 to 1 deployment-controller logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:04 +0000 UTC Normal Pod instance-1-collector-85884cd4-nq4bz AddedInterface Add eth0 [10.129.2.17/23] from ovn-kubernetes multus logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:04 +0000 UTC Normal Pod instance-1-collector-85884cd4-nq4bz.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:04 +0000 UTC Normal Pod instance-1-query-8db556c94-98nsv Binding Scheduled Successfully assigned kuttl-test-meet-fox/instance-1-query-8db556c94-98nsv to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:04 +0000 UTC Normal Pod instance-1-query-8db556c94-98nsv AddedInterface Add eth0 [10.128.2.19/23] from ovn-kubernetes multus logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:04 +0000 UTC Normal Pod instance-1-query-8db556c94-98nsv.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:04 +0000 UTC Normal ReplicaSet.apps instance-1-query-8db556c94 SuccessfulCreate Created pod: instance-1-query-8db556c94-98nsv replicaset-controller logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:07 +0000 UTC Normal Pod instance-1-collector-85884cd4-nq4bz.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" in 2.88s (2.88s including waiting) kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:07 +0000 UTC Normal Pod instance-1-collector-85884cd4-nq4bz.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:07 +0000 UTC Normal Pod instance-1-collector-85884cd4-nq4bz.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:07 +0000 UTC Normal Pod instance-1-query-8db556c94-98nsv.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" in 3.087s (3.087s including waiting) kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:07 +0000 UTC Normal Pod instance-1-query-8db556c94-98nsv.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:07 +0000 UTC Normal Pod instance-1-query-8db556c94-98nsv.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:07 +0000 UTC Normal Pod instance-1-query-8db556c94-98nsv.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:07 +0000 UTC Normal Pod instance-1-query-8db556c94-98nsv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:07 +0000 UTC Normal Pod instance-1-query-8db556c94-98nsv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:07 +0000 UTC Normal Pod instance-1-query-8db556c94-98nsv.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:10 +0000 UTC Normal Pod instance-1-query-8db556c94-98nsv.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" in 2.728s (2.728s including waiting) kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:10 +0000 UTC Normal Pod instance-1-query-8db556c94-98nsv.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:10 +0000 UTC Normal Pod instance-1-query-8db556c94-98nsv.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:28 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:28 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod instance-1-collector-85884cd4-nq4bz horizontal-pod-autoscaler logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:28 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod instance-1-collector-85884cd4-nq4bz horizontal-pod-autoscaler logger.go:42: 07:00:08 | es-multiinstance | 2024-06-10 06:59:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod instance-1-collector-85884cd4-nq4bz horizontal-pod-autoscaler logger.go:42: 07:00:08 | es-multiinstance | Deleting namespace: kuttl-test-meet-fox === CONT kuttl/harness/es-simple-prod logger.go:42: 07:00:15 | es-simple-prod | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:00:15 | es-simple-prod | Creating namespace: kuttl-test-honest-sailfish logger.go:42: 07:00:15 | es-simple-prod | es-simple-prod events from ns kuttl-test-honest-sailfish: logger.go:42: 07:00:15 | es-simple-prod | Deleting namespace: kuttl-test-honest-sailfish === CONT kuttl/harness/es-rollover-autoprov logger.go:42: 07:00:21 | es-rollover-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:00:21 | es-rollover-autoprov | Creating namespace: kuttl-test-grand-tetra logger.go:42: 07:00:21 | es-rollover-autoprov/1-install | starting test step 1-install logger.go:42: 07:00:21 | es-rollover-autoprov/1-install | Jaeger:kuttl-test-grand-tetra/my-jaeger created logger.go:42: 07:00:58 | es-rollover-autoprov/1-install | test step completed 1-install logger.go:42: 07:00:58 | es-rollover-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:00:58 | es-rollover-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:01:02 | es-rollover-autoprov/2-report-spans | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:01:12 | es-rollover-autoprov/2-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:01:13 | es-rollover-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:01:14 | es-rollover-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:01:40 | es-rollover-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 07:01:40 | es-rollover-autoprov/3-check-indices | starting test step 3-check-indices logger.go:42: 07:01:40 | es-rollover-autoprov/3-check-indices | Job:kuttl-test-grand-tetra/00-check-indices created logger.go:42: 07:01:45 | es-rollover-autoprov/3-check-indices | test step completed 3-check-indices logger.go:42: 07:01:45 | es-rollover-autoprov/4-check-indices | starting test step 4-check-indices logger.go:42: 07:01:45 | es-rollover-autoprov/4-check-indices | Job:kuttl-test-grand-tetra/01-check-indices created logger.go:42: 07:01:49 | es-rollover-autoprov/4-check-indices | test step completed 4-check-indices logger.go:42: 07:01:49 | es-rollover-autoprov/5-install | starting test step 5-install logger.go:42: 07:01:49 | es-rollover-autoprov/5-install | Jaeger:kuttl-test-grand-tetra/my-jaeger updated logger.go:42: 07:02:03 | es-rollover-autoprov/5-install | test step completed 5-install logger.go:42: 07:02:03 | es-rollover-autoprov/6-report-spans | starting test step 6-report-spans logger.go:42: 07:02:03 | es-rollover-autoprov/6-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:02:11 | es-rollover-autoprov/6-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=02 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-02-job.yaml] logger.go:42: 07:02:12 | es-rollover-autoprov/6-report-spans | running command: [sh -c kubectl apply -f report-span-02-job.yaml -n $NAMESPACE] logger.go:42: 07:02:12 | es-rollover-autoprov/6-report-spans | job.batch/02-report-span created logger.go:42: 07:02:38 | es-rollover-autoprov/6-report-spans | test step completed 6-report-spans logger.go:42: 07:02:38 | es-rollover-autoprov/7-check-indices | starting test step 7-check-indices logger.go:42: 07:02:38 | es-rollover-autoprov/7-check-indices | Job:kuttl-test-grand-tetra/02-check-indices created logger.go:42: 07:02:42 | es-rollover-autoprov/7-check-indices | test step completed 7-check-indices logger.go:42: 07:02:42 | es-rollover-autoprov/8-check-indices | starting test step 8-check-indices logger.go:42: 07:02:42 | es-rollover-autoprov/8-check-indices | Job:kuttl-test-grand-tetra/03-check-indices created logger.go:42: 07:02:46 | es-rollover-autoprov/8-check-indices | test step completed 8-check-indices logger.go:42: 07:02:46 | es-rollover-autoprov/9-check-indices | starting test step 9-check-indices logger.go:42: 07:02:46 | es-rollover-autoprov/9-check-indices | Job:kuttl-test-grand-tetra/04-check-indices created logger.go:42: 07:02:51 | es-rollover-autoprov/9-check-indices | test step completed 9-check-indices logger.go:42: 07:02:51 | es-rollover-autoprov/10-report-spans | starting test step 10-report-spans logger.go:42: 07:02:51 | es-rollover-autoprov/10-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:03:00 | es-rollover-autoprov/10-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=03 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-03-job.yaml] logger.go:42: 07:03:00 | es-rollover-autoprov/10-report-spans | running command: [sh -c kubectl apply -f report-span-03-job.yaml -n $NAMESPACE] logger.go:42: 07:03:02 | es-rollover-autoprov/10-report-spans | job.batch/03-report-span created logger.go:42: 07:03:26 | es-rollover-autoprov/10-report-spans | test step completed 10-report-spans logger.go:42: 07:03:26 | es-rollover-autoprov/11-check-indices | starting test step 11-check-indices logger.go:42: 07:03:26 | es-rollover-autoprov/11-check-indices | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob my-jaeger-es-rollover --namespace $NAMESPACE] logger.go:42: 07:03:43 | es-rollover-autoprov/11-check-indices | time="2024-06-10T07:03:43Z" level=debug msg="Checking if the my-jaeger-es-rollover CronJob exists" logger.go:42: 07:03:43 | es-rollover-autoprov/11-check-indices | time="2024-06-10T07:03:43Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 07:03:43 | es-rollover-autoprov/11-check-indices | time="2024-06-10T07:03:43Z" level=info msg="Cronjob my-jaeger-es-rollover found successfully" logger.go:42: 07:03:43 | es-rollover-autoprov/11-check-indices | time="2024-06-10T07:03:43Z" level=debug msg="Waiting for the next scheduled job from my-jaeger-es-rollover cronjob" logger.go:42: 07:03:43 | es-rollover-autoprov/11-check-indices | time="2024-06-10T07:03:43Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:03:53 | es-rollover-autoprov/11-check-indices | time="2024-06-10T07:03:53Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:04:03 | es-rollover-autoprov/11-check-indices | time="2024-06-10T07:04:03Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:04:13 | es-rollover-autoprov/11-check-indices | time="2024-06-10T07:04:13Z" level=info msg="Job of owner my-jaeger-es-rollover succeeded after my-jaeger-es-rollover 30.050532594s" logger.go:42: 07:04:14 | es-rollover-autoprov/11-check-indices | Job:kuttl-test-grand-tetra/05-check-indices created logger.go:42: 07:04:19 | es-rollover-autoprov/11-check-indices | test step completed 11-check-indices logger.go:42: 07:04:19 | es-rollover-autoprov/12-check-indices | starting test step 12-check-indices logger.go:42: 07:04:19 | es-rollover-autoprov/12-check-indices | Job:kuttl-test-grand-tetra/06-check-indices created logger.go:42: 07:04:23 | es-rollover-autoprov/12-check-indices | test step completed 12-check-indices logger.go:42: 07:04:23 | es-rollover-autoprov | es-rollover-autoprov events from ns kuttl-test-grand-tetra: logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:28 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestgrandtetramyjaeger-1-6b79477c96 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestgrandtetramyjaeger-1-6b79477c967qwkk replicaset-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgrandtetramyjaeger-1-6b79477c967qwkk Binding Scheduled Successfully assigned kuttl-test-grand-tetra/elasticsearch-cdm-kuttltestgrandtetramyjaeger-1-6b79477c967qwkk to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgrandtetramyjaeger-1-6b79477c967qwkk AddedInterface Add eth0 [10.131.0.22/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgrandtetramyjaeger-1-6b79477c967qwkk.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:28 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestgrandtetramyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestgrandtetramyjaeger-1-6b79477c96 to 1 deployment-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgrandtetramyjaeger-1-6b79477c967qwkk.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgrandtetramyjaeger-1-6b79477c967qwkk.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgrandtetramyjaeger-1-6b79477c967qwkk.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgrandtetramyjaeger-1-6b79477c967qwkk.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgrandtetramyjaeger-1-6b79477c967qwkk.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:43 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestgrandtetramyjaeger-1-6b79477c967qwkk.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:55 +0000 UTC Normal Pod my-jaeger-collector-78698fc768-c42w7 Binding Scheduled Successfully assigned kuttl-test-grand-tetra/my-jaeger-collector-78698fc768-c42w7 to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:55 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-78698fc768 SuccessfulCreate Created pod: my-jaeger-collector-78698fc768-c42w7 replicaset-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:55 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-78698fc768 to 1 deployment-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:55 +0000 UTC Normal Pod my-jaeger-query-78c644cc78-45b2s Binding Scheduled Successfully assigned kuttl-test-grand-tetra/my-jaeger-query-78c644cc78-45b2s to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:55 +0000 UTC Normal Pod my-jaeger-query-78c644cc78-45b2s AddedInterface Add eth0 [10.128.2.24/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:55 +0000 UTC Normal Pod my-jaeger-query-78c644cc78-45b2s.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:55 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-78c644cc78 SuccessfulCreate Created pod: my-jaeger-query-78c644cc78-45b2s replicaset-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:55 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-78c644cc78 to 1 deployment-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:56 +0000 UTC Normal Pod my-jaeger-collector-78698fc768-c42w7 AddedInterface Add eth0 [10.128.2.23/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:56 +0000 UTC Normal Pod my-jaeger-collector-78698fc768-c42w7.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" already present on machine kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:56 +0000 UTC Normal Pod my-jaeger-collector-78698fc768-c42w7.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:56 +0000 UTC Normal Pod my-jaeger-collector-78698fc768-c42w7.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:56 +0000 UTC Normal Pod my-jaeger-query-78c644cc78-45b2s.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:56 +0000 UTC Normal Pod my-jaeger-query-78c644cc78-45b2s.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:56 +0000 UTC Normal Pod my-jaeger-query-78c644cc78-45b2s.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:56 +0000 UTC Normal Pod my-jaeger-query-78c644cc78-45b2s.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:56 +0000 UTC Normal Pod my-jaeger-query-78c644cc78-45b2s.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:56 +0000 UTC Normal Pod my-jaeger-query-78c644cc78-45b2s.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:56 +0000 UTC Normal Pod my-jaeger-query-78c644cc78-45b2s.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:00:56 +0000 UTC Normal Pod my-jaeger-query-78c644cc78-45b2s.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:03 +0000 UTC Normal Pod my-jaeger-query-78c644cc78-45b2s.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:03 +0000 UTC Normal Pod my-jaeger-query-78c644cc78-45b2s.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:03 +0000 UTC Normal Pod my-jaeger-query-78c644cc78-45b2s.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:03 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-78c644cc78 SuccessfulDelete Deleted pod: my-jaeger-query-78c644cc78-45b2s replicaset-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:03 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-78c644cc78 to 0 from 1 deployment-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:04 +0000 UTC Normal Pod my-jaeger-query-59c55dcf49-rdlcv Binding Scheduled Successfully assigned kuttl-test-grand-tetra/my-jaeger-query-59c55dcf49-rdlcv to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:04 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-59c55dcf49 SuccessfulCreate Created pod: my-jaeger-query-59c55dcf49-rdlcv replicaset-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:04 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-59c55dcf49 to 1 deployment-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:05 +0000 UTC Normal Pod my-jaeger-query-59c55dcf49-rdlcv AddedInterface Add eth0 [10.129.2.19/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:05 +0000 UTC Normal Pod my-jaeger-query-59c55dcf49-rdlcv.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:07 +0000 UTC Normal Pod my-jaeger-query-59c55dcf49-rdlcv.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" in 2.351s (2.351s including waiting) kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:07 +0000 UTC Normal Pod my-jaeger-query-59c55dcf49-rdlcv.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:07 +0000 UTC Normal Pod my-jaeger-query-59c55dcf49-rdlcv.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:07 +0000 UTC Normal Pod my-jaeger-query-59c55dcf49-rdlcv.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:07 +0000 UTC Normal Pod my-jaeger-query-59c55dcf49-rdlcv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:07 +0000 UTC Normal Pod my-jaeger-query-59c55dcf49-rdlcv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:07 +0000 UTC Normal Pod my-jaeger-query-59c55dcf49-rdlcv.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:11 +0000 UTC Normal Pod my-jaeger-query-59c55dcf49-rdlcv.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" in 3.121s (3.121s including waiting) kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:11 +0000 UTC Normal Pod my-jaeger-query-59c55dcf49-rdlcv.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:11 +0000 UTC Normal Pod my-jaeger-query-59c55dcf49-rdlcv.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:14 +0000 UTC Normal Pod 00-report-span-k66zg Binding Scheduled Successfully assigned kuttl-test-grand-tetra/00-report-span-k66zg to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:14 +0000 UTC Normal Pod 00-report-span-k66zg AddedInterface Add eth0 [10.128.2.25/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:14 +0000 UTC Normal Pod 00-report-span-k66zg.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:14 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-k66zg job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:17 +0000 UTC Normal Pod 00-report-span-k66zg.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.802s (2.802s including waiting) kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:17 +0000 UTC Normal Pod 00-report-span-k66zg.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:17 +0000 UTC Normal Pod 00-report-span-k66zg.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-78698fc768-c42w7 horizontal-pod-autoscaler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:40 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-78698fc768-c42w7 horizontal-pod-autoscaler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-78698fc768-c42w7 horizontal-pod-autoscaler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:41 +0000 UTC Normal Pod 00-check-indices-rwndf Binding Scheduled Successfully assigned kuttl-test-grand-tetra/00-check-indices-rwndf to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:41 +0000 UTC Normal Pod 00-check-indices-rwndf AddedInterface Add eth0 [10.128.2.26/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:41 +0000 UTC Normal Pod 00-check-indices-rwndf.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:41 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-rwndf job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:42 +0000 UTC Normal Pod 00-check-indices-rwndf.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 370ms (370ms including waiting) kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:42 +0000 UTC Normal Pod 00-check-indices-rwndf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:42 +0000 UTC Normal Pod 00-check-indices-rwndf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:44 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:45 +0000 UTC Normal Pod 01-check-indices-gxzjz Binding Scheduled Successfully assigned kuttl-test-grand-tetra/01-check-indices-gxzjz to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:45 +0000 UTC Normal Pod 01-check-indices-gxzjz AddedInterface Add eth0 [10.128.2.27/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:45 +0000 UTC Normal Pod 01-check-indices-gxzjz.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:45 +0000 UTC Normal Job.batch 01-check-indices SuccessfulCreate Created pod: 01-check-indices-gxzjz job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:46 +0000 UTC Normal Pod 01-check-indices-gxzjz.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 343ms (343ms including waiting) kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:46 +0000 UTC Normal Pod 01-check-indices-gxzjz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:46 +0000 UTC Normal Pod 01-check-indices-gxzjz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:48 +0000 UTC Normal Job.batch 01-check-indices Completed Job completed job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:50 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-2g7vw Binding Scheduled Successfully assigned kuttl-test-grand-tetra/my-jaeger-es-rollover-create-mapping-2g7vw to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:50 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping SuccessfulCreate Created pod: my-jaeger-es-rollover-create-mapping-2g7vw job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:51 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-2g7vw AddedInterface Add eth0 [10.128.2.28/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:51 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-2g7vw.spec.containers{my-jaeger-es-rollover-create-mapping} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:caa92aa8c84ecb5e1740a4d41bd4204f134a6a0e86d0f1888b9b0bd59f6d4614" kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:55 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-2g7vw.spec.containers{my-jaeger-es-rollover-create-mapping} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:caa92aa8c84ecb5e1740a4d41bd4204f134a6a0e86d0f1888b9b0bd59f6d4614" in 3.814s (3.814s including waiting) kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:55 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-2g7vw.spec.containers{my-jaeger-es-rollover-create-mapping} Created Created container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:55 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-2g7vw.spec.containers{my-jaeger-es-rollover-create-mapping} Started Started container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:57 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping Completed Job completed job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:58 +0000 UTC Normal Pod my-jaeger-collector-78698fc768-c42w7.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:58 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-78698fc768 SuccessfulDelete Deleted pod: my-jaeger-collector-78698fc768-c42w7 replicaset-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:58 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-78698fc768 to 0 from 1 deployment-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:58 +0000 UTC Normal Pod my-jaeger-query-59c55dcf49-rdlcv.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:58 +0000 UTC Normal Pod my-jaeger-query-59c55dcf49-rdlcv.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:58 +0000 UTC Normal Pod my-jaeger-query-59c55dcf49-rdlcv.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:58 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-59c55dcf49 SuccessfulDelete Deleted pod: my-jaeger-query-59c55dcf49-rdlcv replicaset-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:58 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-59c55dcf49 to 0 from 1 deployment-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:59 +0000 UTC Normal Pod my-jaeger-collector-7bb76fb54c-jvfzb Binding Scheduled Successfully assigned kuttl-test-grand-tetra/my-jaeger-collector-7bb76fb54c-jvfzb to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:59 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7bb76fb54c SuccessfulCreate Created pod: my-jaeger-collector-7bb76fb54c-jvfzb replicaset-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:59 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7bb76fb54c to 1 deployment-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:59 +0000 UTC Normal Pod my-jaeger-query-56dbbfb9f6-8bzrt Binding Scheduled Successfully assigned kuttl-test-grand-tetra/my-jaeger-query-56dbbfb9f6-8bzrt to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:59 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-56dbbfb9f6 SuccessfulCreate Created pod: my-jaeger-query-56dbbfb9f6-8bzrt replicaset-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:01:59 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-56dbbfb9f6 to 1 deployment-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-collector-7bb76fb54c-jvfzb AddedInterface Add eth0 [10.129.2.20/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-collector-7bb76fb54c-jvfzb.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" already present on machine kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-collector-7bb76fb54c-jvfzb.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-collector-7bb76fb54c-jvfzb.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28633382-grbkm Binding Scheduled Successfully assigned kuttl-test-grand-tetra/my-jaeger-es-lookback-28633382-grbkm to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28633382-grbkm AddedInterface Add eth0 [10.129.2.21/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28633382-grbkm.spec.containers{my-jaeger-es-lookback} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:caa92aa8c84ecb5e1740a4d41bd4204f134a6a0e86d0f1888b9b0bd59f6d4614" kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28633382 SuccessfulCreate Created pod: my-jaeger-es-lookback-28633382-grbkm job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28633382 cronjob-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28633382-kt9vv Binding Scheduled Successfully assigned kuttl-test-grand-tetra/my-jaeger-es-rollover-28633382-kt9vv to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28633382-kt9vv AddedInterface Add eth0 [10.129.2.22/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28633382-kt9vv.spec.containers{my-jaeger-es-rollover} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:caa92aa8c84ecb5e1740a4d41bd4204f134a6a0e86d0f1888b9b0bd59f6d4614" kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28633382 SuccessfulCreate Created pod: my-jaeger-es-rollover-28633382-kt9vv job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28633382 cronjob-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-query-56dbbfb9f6-8bzrt AddedInterface Add eth0 [10.128.2.29/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-query-56dbbfb9f6-8bzrt.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-query-56dbbfb9f6-8bzrt.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-query-56dbbfb9f6-8bzrt.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-query-56dbbfb9f6-8bzrt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-query-56dbbfb9f6-8bzrt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-query-56dbbfb9f6-8bzrt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-query-56dbbfb9f6-8bzrt.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-query-56dbbfb9f6-8bzrt.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:00 +0000 UTC Normal Pod my-jaeger-query-56dbbfb9f6-8bzrt.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:02 +0000 UTC Normal Pod my-jaeger-es-lookback-28633382-grbkm.spec.containers{my-jaeger-es-lookback} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:caa92aa8c84ecb5e1740a4d41bd4204f134a6a0e86d0f1888b9b0bd59f6d4614" in 2.163s (2.163s including waiting) kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:03 +0000 UTC Normal Pod my-jaeger-es-lookback-28633382-grbkm.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:03 +0000 UTC Normal Pod my-jaeger-es-lookback-28633382-grbkm.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:03 +0000 UTC Normal Pod my-jaeger-es-rollover-28633382-kt9vv.spec.containers{my-jaeger-es-rollover} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:caa92aa8c84ecb5e1740a4d41bd4204f134a6a0e86d0f1888b9b0bd59f6d4614" in 2.206s (2.206s including waiting) kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:03 +0000 UTC Normal Pod my-jaeger-es-rollover-28633382-kt9vv.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:03 +0000 UTC Normal Pod my-jaeger-es-rollover-28633382-kt9vv.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:06 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28633382 Completed Job completed job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:06 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28633382, status: Complete cronjob-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:06 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28633382 Completed Job completed job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:06 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28633382, status: Complete cronjob-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:12 +0000 UTC Normal Pod 02-report-span-txlh9 Binding Scheduled Successfully assigned kuttl-test-grand-tetra/02-report-span-txlh9 to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:12 +0000 UTC Normal Job.batch 02-report-span SuccessfulCreate Created pod: 02-report-span-txlh9 job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:13 +0000 UTC Normal Pod 02-report-span-txlh9 AddedInterface Add eth0 [10.129.2.23/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:13 +0000 UTC Normal Pod 02-report-span-txlh9.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:14 +0000 UTC Normal Pod 02-report-span-txlh9.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.605s (1.605s including waiting) kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:14 +0000 UTC Normal Pod 02-report-span-txlh9.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:14 +0000 UTC Normal Pod 02-report-span-txlh9.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:37 +0000 UTC Normal Job.batch 02-report-span Completed Job completed job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:38 +0000 UTC Normal Pod 02-check-indices-76gf7 Binding Scheduled Successfully assigned kuttl-test-grand-tetra/02-check-indices-76gf7 to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:38 +0000 UTC Normal Job.batch 02-check-indices SuccessfulCreate Created pod: 02-check-indices-76gf7 job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:39 +0000 UTC Normal Pod 02-check-indices-76gf7 AddedInterface Add eth0 [10.129.2.24/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:39 +0000 UTC Normal Pod 02-check-indices-76gf7.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:39 +0000 UTC Normal Pod 02-check-indices-76gf7.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 358ms (358ms including waiting) kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:39 +0000 UTC Normal Pod 02-check-indices-76gf7.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:39 +0000 UTC Normal Pod 02-check-indices-76gf7.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:42 +0000 UTC Normal Job.batch 02-check-indices Completed Job completed job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:42 +0000 UTC Normal Pod 03-check-indices-xphks Binding Scheduled Successfully assigned kuttl-test-grand-tetra/03-check-indices-xphks to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:42 +0000 UTC Normal Job.batch 03-check-indices SuccessfulCreate Created pod: 03-check-indices-xphks job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:43 +0000 UTC Normal Pod 03-check-indices-xphks AddedInterface Add eth0 [10.129.2.25/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:43 +0000 UTC Normal Pod 03-check-indices-xphks.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:43 +0000 UTC Normal Pod 03-check-indices-xphks.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 333ms (333ms including waiting) kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:43 +0000 UTC Normal Pod 03-check-indices-xphks.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:43 +0000 UTC Normal Pod 03-check-indices-xphks.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:46 +0000 UTC Normal Job.batch 03-check-indices Completed Job completed job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:46 +0000 UTC Normal Pod 04-check-indices-hplq8 Binding Scheduled Successfully assigned kuttl-test-grand-tetra/04-check-indices-hplq8 to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:46 +0000 UTC Normal Job.batch 04-check-indices SuccessfulCreate Created pod: 04-check-indices-hplq8 job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:47 +0000 UTC Normal Pod 04-check-indices-hplq8 AddedInterface Add eth0 [10.129.2.26/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:47 +0000 UTC Normal Pod 04-check-indices-hplq8.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:47 +0000 UTC Normal Pod 04-check-indices-hplq8.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 388ms (388ms including waiting) kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:47 +0000 UTC Normal Pod 04-check-indices-hplq8.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:47 +0000 UTC Normal Pod 04-check-indices-hplq8.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:50 +0000 UTC Normal Job.batch 04-check-indices Completed Job completed job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:55 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-7bb76fb54c-jvfzb horizontal-pod-autoscaler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:55 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-7bb76fb54c-jvfzb horizontal-pod-autoscaler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:02:55 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-7bb76fb54c-jvfzb horizontal-pod-autoscaler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28633383-zd2qx Binding Scheduled Successfully assigned kuttl-test-grand-tetra/my-jaeger-es-lookback-28633383-zd2qx to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28633383-zd2qx AddedInterface Add eth0 [10.129.2.27/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28633383-zd2qx.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:caa92aa8c84ecb5e1740a4d41bd4204f134a6a0e86d0f1888b9b0bd59f6d4614" already present on machine kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28633383-zd2qx.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28633383-zd2qx.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28633383 SuccessfulCreate Created pod: my-jaeger-es-lookback-28633383-zd2qx job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28633383 cronjob-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28633383-qqxxn Binding Scheduled Successfully assigned kuttl-test-grand-tetra/my-jaeger-es-rollover-28633383-qqxxn to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28633383-qqxxn AddedInterface Add eth0 [10.129.2.28/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28633383-qqxxn.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:caa92aa8c84ecb5e1740a4d41bd4204f134a6a0e86d0f1888b9b0bd59f6d4614" already present on machine kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28633383-qqxxn.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28633383-qqxxn.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28633383 SuccessfulCreate Created pod: my-jaeger-es-rollover-28633383-qqxxn job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28633383 cronjob-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:02 +0000 UTC Normal Pod 03-report-span-47q5l Binding Scheduled Successfully assigned kuttl-test-grand-tetra/03-report-span-47q5l to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:02 +0000 UTC Normal Job.batch 03-report-span SuccessfulCreate Created pod: 03-report-span-47q5l job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:03 +0000 UTC Normal Pod 03-report-span-47q5l AddedInterface Add eth0 [10.128.2.30/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:03 +0000 UTC Normal Pod 03-report-span-47q5l.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:03 +0000 UTC Normal Pod 03-report-span-47q5l.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 391ms (391ms including waiting) kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:03 +0000 UTC Normal Pod 03-report-span-47q5l.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:03 +0000 UTC Normal Pod 03-report-span-47q5l.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28633383 Completed Job completed job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28633383, status: Complete cronjob-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:04 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28633383 Completed Job completed job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:04 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28633383, status: Complete cronjob-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:03:26 +0000 UTC Normal Job.batch 03-report-span Completed Job completed job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28633384-bsxkd Binding Scheduled Successfully assigned kuttl-test-grand-tetra/my-jaeger-es-lookback-28633384-bsxkd to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28633384-bsxkd AddedInterface Add eth0 [10.129.2.30/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28633384-bsxkd.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:caa92aa8c84ecb5e1740a4d41bd4204f134a6a0e86d0f1888b9b0bd59f6d4614" already present on machine kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28633384-bsxkd.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28633384-bsxkd.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28633384 SuccessfulCreate Created pod: my-jaeger-es-lookback-28633384-bsxkd job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28633384 cronjob-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28633384-54g9z Binding Scheduled Successfully assigned kuttl-test-grand-tetra/my-jaeger-es-rollover-28633384-54g9z to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28633384 SuccessfulCreate Created pod: my-jaeger-es-rollover-28633384-54g9z job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28633384 cronjob-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:01 +0000 UTC Normal Pod my-jaeger-es-rollover-28633384-54g9z AddedInterface Add eth0 [10.129.2.29/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:01 +0000 UTC Normal Pod my-jaeger-es-rollover-28633384-54g9z.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:caa92aa8c84ecb5e1740a4d41bd4204f134a6a0e86d0f1888b9b0bd59f6d4614" already present on machine kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:01 +0000 UTC Normal Pod my-jaeger-es-rollover-28633384-54g9z.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:01 +0000 UTC Normal Pod my-jaeger-es-rollover-28633384-54g9z.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28633384 Completed Job completed job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28633384, status: Complete cronjob-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:04 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28633384 Completed Job completed job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:04 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28633384, status: Complete cronjob-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:14 +0000 UTC Normal Pod 05-check-indices-5bvsm Binding Scheduled Successfully assigned kuttl-test-grand-tetra/05-check-indices-5bvsm to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:14 +0000 UTC Normal Pod 05-check-indices-5bvsm AddedInterface Add eth0 [10.129.2.31/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:14 +0000 UTC Normal Pod 05-check-indices-5bvsm.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:14 +0000 UTC Normal Job.batch 05-check-indices SuccessfulCreate Created pod: 05-check-indices-5bvsm job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:15 +0000 UTC Normal Pod 05-check-indices-5bvsm.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 381ms (381ms including waiting) kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:15 +0000 UTC Normal Pod 05-check-indices-5bvsm.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:15 +0000 UTC Normal Pod 05-check-indices-5bvsm.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:18 +0000 UTC Normal Job.batch 05-check-indices Completed Job completed job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:19 +0000 UTC Normal Pod 06-check-indices-rf8nc Binding Scheduled Successfully assigned kuttl-test-grand-tetra/06-check-indices-rf8nc to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:19 +0000 UTC Normal Job.batch 06-check-indices SuccessfulCreate Created pod: 06-check-indices-rf8nc job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:20 +0000 UTC Normal Pod 06-check-indices-rf8nc AddedInterface Add eth0 [10.129.2.32/23] from ovn-kubernetes multus logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:20 +0000 UTC Normal Pod 06-check-indices-rf8nc.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:20 +0000 UTC Normal Pod 06-check-indices-rf8nc.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 322ms (322ms including waiting) kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:20 +0000 UTC Normal Pod 06-check-indices-rf8nc.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:20 +0000 UTC Normal Pod 06-check-indices-rf8nc.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:04:23 | es-rollover-autoprov | 2024-06-10 07:04:23 +0000 UTC Normal Job.batch 06-check-indices Completed Job completed job-controller logger.go:42: 07:04:23 | es-rollover-autoprov | Deleting namespace: kuttl-test-grand-tetra === CONT kuttl/harness/es-increasing-replicas logger.go:42: 07:04:31 | es-increasing-replicas | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:04:31 | es-increasing-replicas | Ignoring check-es-nodes.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:04:31 | es-increasing-replicas | Ignoring openshift-check-es-nodes.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:04:31 | es-increasing-replicas | Creating namespace: kuttl-test-dynamic-husky logger.go:42: 07:04:31 | es-increasing-replicas/1-install | starting test step 1-install logger.go:42: 07:04:31 | es-increasing-replicas/1-install | Jaeger:kuttl-test-dynamic-husky/simple-prod created logger.go:42: 07:05:06 | es-increasing-replicas/1-install | test step completed 1-install logger.go:42: 07:05:06 | es-increasing-replicas/2-install | starting test step 2-install logger.go:42: 07:05:06 | es-increasing-replicas/2-install | Jaeger:kuttl-test-dynamic-husky/simple-prod updated logger.go:42: 07:05:20 | es-increasing-replicas/2-install | test step completed 2-install logger.go:42: 07:05:20 | es-increasing-replicas/3-smoke-test | starting test step 3-smoke-test logger.go:42: 07:05:20 | es-increasing-replicas/3-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:05:21 | es-increasing-replicas/3-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:05:28 | es-increasing-replicas/3-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:05:28 | es-increasing-replicas/3-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:05:29 | es-increasing-replicas/3-smoke-test | job.batch/report-span created logger.go:42: 07:05:29 | es-increasing-replicas/3-smoke-test | job.batch/check-span created logger.go:42: 07:05:41 | es-increasing-replicas/3-smoke-test | test step completed 3-smoke-test logger.go:42: 07:05:41 | es-increasing-replicas/4-install | starting test step 4-install logger.go:42: 07:05:41 | es-increasing-replicas/4-install | Jaeger:kuttl-test-dynamic-husky/simple-prod updated logger.go:42: 07:05:41 | es-increasing-replicas/4-install | test step completed 4-install logger.go:42: 07:05:41 | es-increasing-replicas/5-check-es-nodes | starting test step 5-check-es-nodes logger.go:42: 07:05:41 | es-increasing-replicas/5-check-es-nodes | running command: [sh -c ./check-es-nodes.sh $NAMESPACE] logger.go:42: 07:05:41 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:05:41 | es-increasing-replicas/5-check-es-nodes | false logger.go:42: 07:05:41 | es-increasing-replicas/5-check-es-nodes | Error: no matches found logger.go:42: 07:05:46 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:05:47 | es-increasing-replicas/5-check-es-nodes | true logger.go:42: 07:05:47 | es-increasing-replicas/5-check-es-nodes | test step completed 5-check-es-nodes logger.go:42: 07:05:47 | es-increasing-replicas | es-increasing-replicas events from ns kuttl-test-dynamic-husky: logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:04:37 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestdynamichuskysimpleprod-1-58558565b4 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestdynamichuskysimpleprod-1-585585bfgdq replicaset-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:04:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdynamichuskysimpleprod-1-585585bfgdq Binding Scheduled Successfully assigned kuttl-test-dynamic-husky/elasticsearch-cdm-kuttltestdynamichuskysimpleprod-1-585585bfgdq to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:04:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdynamichuskysimpleprod-1-585585bfgdq AddedInterface Add eth0 [10.129.2.33/23] from ovn-kubernetes multus logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:04:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdynamichuskysimpleprod-1-585585bfgdq.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:04:37 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestdynamichuskysimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestdynamichuskysimpleprod-1-58558565b4 to 1 deployment-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:04:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdynamichuskysimpleprod-1-585585bfgdq.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:04:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdynamichuskysimpleprod-1-585585bfgdq.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:04:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdynamichuskysimpleprod-1-585585bfgdq.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:04:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdynamichuskysimpleprod-1-585585bfgdq.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:04:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdynamichuskysimpleprod-1-585585bfgdq.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:04:52 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestdynamichuskysimpleprod-1-585585bfgdq.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:04 +0000 UTC Normal Pod simple-prod-collector-6ddf77486f-mqr8m Binding Scheduled Successfully assigned kuttl-test-dynamic-husky/simple-prod-collector-6ddf77486f-mqr8m to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:04 +0000 UTC Normal Pod simple-prod-collector-6ddf77486f-mqr8m AddedInterface Add eth0 [10.128.2.31/23] from ovn-kubernetes multus logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:04 +0000 UTC Normal Pod simple-prod-collector-6ddf77486f-mqr8m.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:04 +0000 UTC Normal Pod simple-prod-collector-6ddf77486f-mqr8m.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:04 +0000 UTC Normal Pod simple-prod-collector-6ddf77486f-mqr8m.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:04 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-6ddf77486f SuccessfulCreate Created pod: simple-prod-collector-6ddf77486f-mqr8m replicaset-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:04 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-6ddf77486f to 1 deployment-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:04 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-gh4mf Binding Scheduled Successfully assigned kuttl-test-dynamic-husky/simple-prod-query-68dd9cf6d8-gh4mf to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:04 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-gh4mf AddedInterface Add eth0 [10.128.2.32/23] from ovn-kubernetes multus logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:04 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-gh4mf.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:04 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-gh4mf.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:04 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-gh4mf.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:04 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-gh4mf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:04 +0000 UTC Normal ReplicaSet.apps simple-prod-query-68dd9cf6d8 SuccessfulCreate Created pod: simple-prod-query-68dd9cf6d8-gh4mf replicaset-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:04 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-68dd9cf6d8 to 1 deployment-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:05 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-gh4mf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:05 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-gh4mf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:05 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-gh4mf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:05 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-gh4mf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:05 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-gh4mf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:10 +0000 UTC Normal Pod simple-prod-collector-6ddf77486f-sv497 Binding Scheduled Successfully assigned kuttl-test-dynamic-husky/simple-prod-collector-6ddf77486f-sv497 to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:10 +0000 UTC Normal Pod simple-prod-collector-6ddf77486f-sv497 AddedInterface Add eth0 [10.131.0.23/23] from ovn-kubernetes multus logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:10 +0000 UTC Normal Pod simple-prod-collector-6ddf77486f-sv497.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:10 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-6ddf77486f SuccessfulCreate Created pod: simple-prod-collector-6ddf77486f-sv497 replicaset-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:10 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-6ddf77486f to 2 from 1 deployment-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:10 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-9jmvd Binding Scheduled Successfully assigned kuttl-test-dynamic-husky/simple-prod-query-68dd9cf6d8-9jmvd to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:10 +0000 UTC Normal ReplicaSet.apps simple-prod-query-68dd9cf6d8 SuccessfulCreate Created pod: simple-prod-query-68dd9cf6d8-9jmvd replicaset-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:10 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-68dd9cf6d8 to 2 from 1 deployment-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:11 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-9jmvd AddedInterface Add eth0 [10.131.0.24/23] from ovn-kubernetes multus logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:11 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-9jmvd.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:15 +0000 UTC Normal Pod simple-prod-collector-6ddf77486f-sv497.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" in 4.289s (4.289s including waiting) kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:15 +0000 UTC Normal Pod simple-prod-collector-6ddf77486f-sv497.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:15 +0000 UTC Normal Pod simple-prod-collector-6ddf77486f-sv497.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:15 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-9jmvd.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" in 4.238s (4.238s including waiting) kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:15 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-9jmvd.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:15 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-9jmvd.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:15 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-9jmvd.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:15 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-9jmvd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:15 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-9jmvd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:15 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-9jmvd.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:18 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-9jmvd.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" in 2.901s (2.901s including waiting) kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:18 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-9jmvd.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:18 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-9jmvd.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:23 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-9jmvd.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:23 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-9jmvd.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:23 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-9jmvd.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:23 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-gh4mf.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:23 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-gh4mf.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:23 +0000 UTC Normal Pod simple-prod-query-68dd9cf6d8-gh4mf.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:23 +0000 UTC Normal ReplicaSet.apps simple-prod-query-68dd9cf6d8 SuccessfulDelete Deleted pod: simple-prod-query-68dd9cf6d8-9jmvd replicaset-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:23 +0000 UTC Normal ReplicaSet.apps simple-prod-query-68dd9cf6d8 SuccessfulDelete Deleted pod: simple-prod-query-68dd9cf6d8-gh4mf replicaset-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:23 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-68dd9cf6d8 to 0 from 2 deployment-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:24 +0000 UTC Normal Pod simple-prod-query-864f8f854-95zq9 Binding Scheduled Successfully assigned kuttl-test-dynamic-husky/simple-prod-query-864f8f854-95zq9 to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:24 +0000 UTC Normal Pod simple-prod-query-864f8f854-95zq9 AddedInterface Add eth0 [10.131.0.25/23] from ovn-kubernetes multus logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:24 +0000 UTC Normal Pod simple-prod-query-864f8f854-95zq9.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:24 +0000 UTC Normal Pod simple-prod-query-864f8f854-xk27n Binding Scheduled Successfully assigned kuttl-test-dynamic-husky/simple-prod-query-864f8f854-xk27n to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:24 +0000 UTC Normal Pod simple-prod-query-864f8f854-xk27n AddedInterface Add eth0 [10.128.2.33/23] from ovn-kubernetes multus logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:24 +0000 UTC Normal Pod simple-prod-query-864f8f854-xk27n.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:24 +0000 UTC Normal ReplicaSet.apps simple-prod-query-864f8f854 SuccessfulCreate Created pod: simple-prod-query-864f8f854-95zq9 replicaset-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:24 +0000 UTC Normal ReplicaSet.apps simple-prod-query-864f8f854 SuccessfulCreate Created pod: simple-prod-query-864f8f854-xk27n replicaset-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:24 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-864f8f854 to 2 deployment-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:25 +0000 UTC Normal Pod simple-prod-query-864f8f854-95zq9.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:25 +0000 UTC Normal Pod simple-prod-query-864f8f854-95zq9.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:25 +0000 UTC Normal Pod simple-prod-query-864f8f854-95zq9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:25 +0000 UTC Normal Pod simple-prod-query-864f8f854-95zq9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:25 +0000 UTC Normal Pod simple-prod-query-864f8f854-95zq9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:25 +0000 UTC Normal Pod simple-prod-query-864f8f854-95zq9.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:25 +0000 UTC Normal Pod simple-prod-query-864f8f854-95zq9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:25 +0000 UTC Normal Pod simple-prod-query-864f8f854-95zq9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:25 +0000 UTC Normal Pod simple-prod-query-864f8f854-xk27n.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:25 +0000 UTC Normal Pod simple-prod-query-864f8f854-xk27n.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:25 +0000 UTC Normal Pod simple-prod-query-864f8f854-xk27n.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:25 +0000 UTC Normal Pod simple-prod-query-864f8f854-xk27n.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:25 +0000 UTC Normal Pod simple-prod-query-864f8f854-xk27n.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:25 +0000 UTC Normal Pod simple-prod-query-864f8f854-xk27n.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:25 +0000 UTC Normal Pod simple-prod-query-864f8f854-xk27n.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:25 +0000 UTC Normal Pod simple-prod-query-864f8f854-xk27n.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:29 +0000 UTC Normal Pod check-span-z2gk7 Binding Scheduled Successfully assigned kuttl-test-dynamic-husky/check-span-z2gk7 to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:29 +0000 UTC Normal Pod check-span-z2gk7 AddedInterface Add eth0 [10.128.2.34/23] from ovn-kubernetes multus logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:29 +0000 UTC Normal Pod check-span-z2gk7.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:29 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-z2gk7 job-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:29 +0000 UTC Normal Pod report-span-kq6sv Binding Scheduled Successfully assigned kuttl-test-dynamic-husky/report-span-kq6sv to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:29 +0000 UTC Normal Pod report-span-kq6sv AddedInterface Add eth0 [10.131.0.26/23] from ovn-kubernetes multus logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:29 +0000 UTC Normal Pod report-span-kq6sv.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:29 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-kq6sv job-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:30 +0000 UTC Normal Pod check-span-z2gk7.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 318ms (318ms including waiting) kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:30 +0000 UTC Normal Pod check-span-z2gk7.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:30 +0000 UTC Normal Pod check-span-z2gk7.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:31 +0000 UTC Normal Pod report-span-kq6sv.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.677s (1.677s including waiting) kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:31 +0000 UTC Normal Pod report-span-kq6sv.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:31 +0000 UTC Normal Pod report-span-kq6sv.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:41 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:42 +0000 UTC Normal Pod simple-prod-collector-6ddf77486f-sv497.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:42 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-6ddf77486f SuccessfulDelete Deleted pod: simple-prod-collector-6ddf77486f-sv497 replicaset-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:42 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled down replica set simple-prod-collector-6ddf77486f to 0 from 2 deployment-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestdynamichuskysimpleprod-2-d5457f9b6 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestdynamichuskysimpleprod-2-d5457fqjjfk replicaset-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdynamichuskysimpleprod-2-d5457fqjjfk Binding Scheduled Successfully assigned kuttl-test-dynamic-husky/elasticsearch-cdm-kuttltestdynamichuskysimpleprod-2-d5457fqjjfk to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdynamichuskysimpleprod-2-d5457fqjjfk AddedInterface Add eth0 [10.131.0.27/23] from ovn-kubernetes multus logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdynamichuskysimpleprod-2-d5457fqjjfk.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdynamichuskysimpleprod-2-d5457fqjjfk.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdynamichuskysimpleprod-2-d5457fqjjfk.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdynamichuskysimpleprod-2-d5457fqjjfk.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestdynamichuskysimpleprod-2 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestdynamichuskysimpleprod-2-d5457f9b6 to 1 deployment-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal Pod simple-prod-collector-54f485b846-hsnql Binding Scheduled Successfully assigned kuttl-test-dynamic-husky/simple-prod-collector-54f485b846-hsnql to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal Pod simple-prod-collector-54f485b846-t4zhq Binding Scheduled Successfully assigned kuttl-test-dynamic-husky/simple-prod-collector-54f485b846-t4zhq to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-54f485b846 SuccessfulCreate Created pod: simple-prod-collector-54f485b846-t4zhq replicaset-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-54f485b846 SuccessfulCreate Created pod: simple-prod-collector-54f485b846-hsnql replicaset-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal Pod simple-prod-collector-6ddf77486f-mqr8m.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-6ddf77486f SuccessfulDelete Deleted pod: simple-prod-collector-6ddf77486f-mqr8m replicaset-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-54f485b846 to 2 deployment-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal Pod simple-prod-query-864f8f854-95zq9.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal Pod simple-prod-query-864f8f854-95zq9.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal Pod simple-prod-query-864f8f854-95zq9.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal Pod simple-prod-query-864f8f854-xk27n.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal Pod simple-prod-query-864f8f854-xk27n.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal Pod simple-prod-query-864f8f854-xk27n.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal ReplicaSet.apps simple-prod-query-864f8f854 SuccessfulDelete Deleted pod: simple-prod-query-864f8f854-xk27n replicaset-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal ReplicaSet.apps simple-prod-query-864f8f854 SuccessfulDelete Deleted pod: simple-prod-query-864f8f854-95zq9 replicaset-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:43 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-864f8f854 to 0 from 2 deployment-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdynamichuskysimpleprod-2-d5457fqjjfk.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdynamichuskysimpleprod-2-d5457fqjjfk.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:44 +0000 UTC Normal Pod simple-prod-collector-54f485b846-hsnql AddedInterface Add eth0 [10.131.0.28/23] from ovn-kubernetes multus logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:44 +0000 UTC Normal Pod simple-prod-collector-54f485b846-hsnql.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:44 +0000 UTC Normal Pod simple-prod-collector-54f485b846-hsnql.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:44 +0000 UTC Normal Pod simple-prod-collector-54f485b846-t4zhq AddedInterface Add eth0 [10.128.2.35/23] from ovn-kubernetes multus logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:44 +0000 UTC Normal Pod simple-prod-collector-54f485b846-t4zhq.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:44 +0000 UTC Normal Pod simple-prod-collector-54f485b846-t4zhq.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:44 +0000 UTC Normal Pod simple-prod-collector-54f485b846-t4zhq.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:44 +0000 UTC Normal Pod simple-prod-query-b75768b79-ktdvv Binding Scheduled Successfully assigned kuttl-test-dynamic-husky/simple-prod-query-b75768b79-ktdvv to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:44 +0000 UTC Normal Pod simple-prod-query-b75768b79-ktdvv AddedInterface Add eth0 [10.128.2.36/23] from ovn-kubernetes multus logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:44 +0000 UTC Normal Pod simple-prod-query-b75768b79-q2h9r Binding Scheduled Successfully assigned kuttl-test-dynamic-husky/simple-prod-query-b75768b79-q2h9r to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:44 +0000 UTC Normal Pod simple-prod-query-b75768b79-q2h9r AddedInterface Add eth0 [10.131.0.29/23] from ovn-kubernetes multus logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:44 +0000 UTC Normal ReplicaSet.apps simple-prod-query-b75768b79 SuccessfulCreate Created pod: simple-prod-query-b75768b79-ktdvv replicaset-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:44 +0000 UTC Normal ReplicaSet.apps simple-prod-query-b75768b79 SuccessfulCreate Created pod: simple-prod-query-b75768b79-q2h9r replicaset-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:44 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-b75768b79 to 2 deployment-controller logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-collector-54f485b846-hsnql.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-query-b75768b79-ktdvv.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-query-b75768b79-ktdvv.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-query-b75768b79-ktdvv.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-query-b75768b79-ktdvv.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-query-b75768b79-ktdvv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-query-b75768b79-ktdvv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-query-b75768b79-ktdvv.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-query-b75768b79-ktdvv.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-query-b75768b79-ktdvv.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-query-b75768b79-q2h9r.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-query-b75768b79-q2h9r.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-query-b75768b79-q2h9r.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-query-b75768b79-q2h9r.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-query-b75768b79-q2h9r.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-query-b75768b79-q2h9r.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-query-b75768b79-q2h9r.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-query-b75768b79-q2h9r.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:05:47 | es-increasing-replicas | 2024-06-10 07:05:45 +0000 UTC Normal Pod simple-prod-query-b75768b79-q2h9r.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:05:47 | es-increasing-replicas | Deleting namespace: kuttl-test-dynamic-husky === CONT kuttl/harness/es-index-cleaner-autoprov logger.go:42: 07:06:22 | es-index-cleaner-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:06:22 | es-index-cleaner-autoprov | Creating namespace: kuttl-test-meet-hornet logger.go:42: 07:06:22 | es-index-cleaner-autoprov/1-install | starting test step 1-install logger.go:42: 07:06:22 | es-index-cleaner-autoprov/1-install | Jaeger:kuttl-test-meet-hornet/test-es-index-cleaner-with-prefix created logger.go:42: 07:06:59 | es-index-cleaner-autoprov/1-install | test step completed 1-install logger.go:42: 07:06:59 | es-index-cleaner-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:06:59 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE test-es-index-cleaner-with-prefix /dev/null] logger.go:42: 07:07:00 | es-index-cleaner-autoprov/2-report-spans | Warning: resource jaegers/test-es-index-cleaner-with-prefix is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:07:07 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c DAYS=5 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:07:07 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:07:08 | es-index-cleaner-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:17:09 | es-index-cleaner-autoprov/2-report-spans | test step failed 2-report-spans case.go:364: failed in step 2-report-spans case.go:366: --- Job:kuttl-test-meet-hornet/00-report-span +++ Job:kuttl-test-meet-hornet/00-report-span @@ -1,8 +1,162 @@ apiVersion: batch/v1 kind: Job metadata: + annotations: + kubectl.kubernetes.io/last-applied-configuration: | + {"apiVersion":"batch/v1","kind":"Job","metadata":{"annotations":{},"name":"00-report-span","namespace":"kuttl-test-meet-hornet"},"spec":{"backoffLimit":15,"template":{"spec":{"containers":[{"command":["./reporter","--days","5","--verbose"],"env":[{"name":"JAEGER_SERVICE_NAME","value":"smoke-test-service"},{"name":"OPERATION_NAME","value":"smoke-test-operation"},{"name":"JAEGER_ENDPOINT","value":"http://test-es-index-cleaner-with-prefix-collector-headless:14268/api/traces"},{"name":"JAEGER_QUERY","value":"https://test-es-index-cleaner-with-prefix-query/api/traces"},{"name":"SECRET_PATH","value":"/var/run/secrets/api-token/token"}],"image":"quay.io/rhn_support_ikanse/jaeger-asserts:latest","name":"asserts-container","volumeMounts":[{"mountPath":"/var/run/secrets/api-token","name":"token-api-volume"}]}],"restartPolicy":"OnFailure","volumes":[{"name":"token-api-volume","secret":{"secretName":"e2e-test"}}]}}}} + labels: + batch.kubernetes.io/controller-uid: f969ba9a-bd39-42ec-b606-a89b79ca8c30 + batch.kubernetes.io/job-name: 00-report-span + controller-uid: f969ba9a-bd39-42ec-b606-a89b79ca8c30 + job-name: 00-report-span + managedFields: + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:kubectl.kubernetes.io/last-applied-configuration: {} + f:spec: + f:backoffLimit: {} + f:completionMode: {} + f:completions: {} + f:manualSelector: {} + f:parallelism: {} + f:podReplacementPolicy: {} + f:suspend: {} + f:template: + f:spec: + f:containers: + k:{"name":"asserts-container"}: + .: {} + f:command: {} + f:env: + .: {} + k:{"name":"JAEGER_ENDPOINT"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"JAEGER_QUERY"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"JAEGER_SERVICE_NAME"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"OPERATION_NAME"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SECRET_PATH"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:name: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/var/run/secrets/api-token"}: + .: {} + f:mountPath: {} + f:name: {} + f:dnsPolicy: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"token-api-volume"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + manager: kubectl-client-side-apply + operation: Update + time: "2024-06-10T07:07:08Z" + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:status: + f:active: {} + f:ready: {} + f:startTime: {} + f:terminating: {} + f:uncountedTerminatedPods: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-06-10T07:07:11Z" name: 00-report-span namespace: kuttl-test-meet-hornet +spec: + backoffLimit: 15 + completionMode: NonIndexed + completions: 1 + manualSelector: false + parallelism: 1 + podReplacementPolicy: TerminatingOrFailed + selector: + matchLabels: + batch.kubernetes.io/controller-uid: f969ba9a-bd39-42ec-b606-a89b79ca8c30 + suspend: false + template: + metadata: + creationTimestamp: null + labels: + batch.kubernetes.io/controller-uid: f969ba9a-bd39-42ec-b606-a89b79ca8c30 + batch.kubernetes.io/job-name: 00-report-span + controller-uid: f969ba9a-bd39-42ec-b606-a89b79ca8c30 + job-name: 00-report-span + spec: + containers: + - command: + - ./reporter + - --days + - "5" + - --verbose + env: + - name: JAEGER_SERVICE_NAME + value: smoke-test-service + - name: OPERATION_NAME + value: smoke-test-operation + - name: JAEGER_ENDPOINT + value: http://test-es-index-cleaner-with-prefix-collector-headless:14268/api/traces + - name: JAEGER_QUERY + value: https://test-es-index-cleaner-with-prefix-query/api/traces + - name: SECRET_PATH + value: /var/run/secrets/api-token/token + image: quay.io/rhn_support_ikanse/jaeger-asserts:latest + imagePullPolicy: Always + name: asserts-container + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /var/run/secrets/api-token + name: token-api-volume + dnsPolicy: ClusterFirst + restartPolicy: OnFailure + schedulerName: default-scheduler + securityContext: {} + terminationGracePeriodSeconds: 30 + volumes: + - name: token-api-volume + secret: + defaultMode: 420 + secretName: e2e-test status: - succeeded: 1 + active: 1 + ready: 1 + startTime: "2024-06-10T07:07:08Z" + terminating: 0 + uncountedTerminatedPods: {} case.go:366: resource Job:kuttl-test-meet-hornet/00-report-span: .status.succeeded: key is missing from map logger.go:42: 07:17:09 | es-index-cleaner-autoprov | es-index-cleaner-autoprov events from ns kuttl-test-meet-hornet: logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:29 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestmeethornettestesindexcleane-1-54bb58df4 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestmeethornettestesindexcleane-1-59vd7t replicaset-controller logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeethornettestesindexcleane-1-59vd7t Binding Scheduled Successfully assigned kuttl-test-meet-hornet/elasticsearch-cdm-kuttltestmeethornettestesindexcleane-1-59vd7t to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeethornettestesindexcleane-1-59vd7t AddedInterface Add eth0 [10.129.2.34/23] from ovn-kubernetes multus logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeethornettestesindexcleane-1-59vd7t.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeethornettestesindexcleane-1-59vd7t.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeethornettestesindexcleane-1-59vd7t.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeethornettestesindexcleane-1-59vd7t.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeethornettestesindexcleane-1-59vd7t.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmeethornettestesindexcleane-1-59vd7t.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:29 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestmeethornettestesindexcleane-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestmeethornettestesindexcleane-1-54bb58df4 to 1 deployment-controller logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:39 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmeethornettestesindexcleane-1-59vd7t.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:44 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmeethornettestesindexcleane-1-59vd7t.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:55 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-7767799cf5-qx9kf Binding Scheduled Successfully assigned kuttl-test-meet-hornet/test-es-index-cleaner-with-prefix-collector-7767799cf5-qx9kf to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:55 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-collector-7767799cf5 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-collector-7767799cf5-qx9kf replicaset-controller logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:55 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-collector ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-collector-7767799cf5 to 1 deployment-controller logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:55 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5bb9cb7dc5-7962d Binding Scheduled Successfully assigned kuttl-test-meet-hornet/test-es-index-cleaner-with-prefix-query-5bb9cb7dc5-7962d to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:55 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-5bb9cb7dc5 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-5bb9cb7dc5-7962d replicaset-controller logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:55 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-5bb9cb7dc5 to 1 deployment-controller logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:56 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-7767799cf5-qx9kf AddedInterface Add eth0 [10.131.0.30/23] from ovn-kubernetes multus logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:56 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-7767799cf5-qx9kf.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" already present on machine kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:56 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-7767799cf5-qx9kf.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:56 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-7767799cf5-qx9kf.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:56 +0000 UTC Warning Pod test-es-index-cleaner-with-prefix-query-5bb9cb7dc5-7962d FailedMount MountVolume.SetUp failed for volume "test-es-index-cleaner-with-prefix-ui-oauth-proxy-tls" : secret "test-es-index-cleaner-with-prefix-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5bb9cb7dc5-7962d AddedInterface Add eth0 [10.128.2.37/23] from ovn-kubernetes multus logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5bb9cb7dc5-7962d.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5bb9cb7dc5-7962d.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5bb9cb7dc5-7962d.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5bb9cb7dc5-7962d.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5bb9cb7dc5-7962d.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5bb9cb7dc5-7962d.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5bb9cb7dc5-7962d.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5bb9cb7dc5-7962d.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:06:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-5bb9cb7dc5-7962d.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:07:08 +0000 UTC Normal Pod 00-report-span-lh2fv Binding Scheduled Successfully assigned kuttl-test-meet-hornet/00-report-span-lh2fv to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:07:08 +0000 UTC Normal Pod 00-report-span-lh2fv AddedInterface Add eth0 [10.131.0.31/23] from ovn-kubernetes multus logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:07:08 +0000 UTC Normal Pod 00-report-span-lh2fv.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:07:08 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-lh2fv job-controller logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:07:09 +0000 UTC Normal Pod 00-report-span-lh2fv.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 454ms (454ms including waiting) kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:07:09 +0000 UTC Normal Pod 00-report-span-lh2fv.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:07:09 +0000 UTC Normal Pod 00-report-span-lh2fv.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:07:13 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:07:13 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-7767799cf5-qx9kf horizontal-pod-autoscaler logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:07:13 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:07:28 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-7767799cf5-qx9kf horizontal-pod-autoscaler logger.go:42: 07:17:09 | es-index-cleaner-autoprov | 2024-06-10 07:07:28 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-7767799cf5-qx9kf horizontal-pod-autoscaler logger.go:42: 07:17:09 | es-index-cleaner-autoprov | Deleting namespace: kuttl-test-meet-hornet === CONT kuttl/harness/es-from-aio-to-production logger.go:42: 07:17:21 | es-from-aio-to-production | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:17:21 | es-from-aio-to-production | Creating namespace: kuttl-test-sweeping-mallard logger.go:42: 07:17:21 | es-from-aio-to-production/0-install | starting test step 0-install logger.go:42: 07:17:21 | es-from-aio-to-production/0-install | Jaeger:kuttl-test-sweeping-mallard/my-jaeger created logger.go:42: 07:17:29 | es-from-aio-to-production/0-install | test step completed 0-install logger.go:42: 07:17:29 | es-from-aio-to-production/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:17:29 | es-from-aio-to-production/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:17:31 | es-from-aio-to-production/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:17:39 | es-from-aio-to-production/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:17:39 | es-from-aio-to-production/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:17:39 | es-from-aio-to-production/1-smoke-test | job.batch/report-span created logger.go:42: 07:17:39 | es-from-aio-to-production/1-smoke-test | job.batch/check-span created logger.go:42: 07:17:52 | es-from-aio-to-production/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:17:52 | es-from-aio-to-production/3-install | starting test step 3-install logger.go:42: 07:17:52 | es-from-aio-to-production/3-install | Jaeger:kuttl-test-sweeping-mallard/my-jaeger updated logger.go:42: 07:18:25 | es-from-aio-to-production/3-install | test step completed 3-install logger.go:42: 07:18:25 | es-from-aio-to-production/4-smoke-test | starting test step 4-smoke-test logger.go:42: 07:18:25 | es-from-aio-to-production/4-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:18:33 | es-from-aio-to-production/4-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:18:33 | es-from-aio-to-production/4-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:18:34 | es-from-aio-to-production/4-smoke-test | job.batch/report-span unchanged logger.go:42: 07:18:34 | es-from-aio-to-production/4-smoke-test | job.batch/check-span unchanged logger.go:42: 07:18:34 | es-from-aio-to-production/4-smoke-test | test step completed 4-smoke-test logger.go:42: 07:18:34 | es-from-aio-to-production | es-from-aio-to-production events from ns kuttl-test-sweeping-mallard: logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:25 +0000 UTC Normal Pod my-jaeger-b7b4b4b99-54rs6 Binding Scheduled Successfully assigned kuttl-test-sweeping-mallard/my-jaeger-b7b4b4b99-54rs6 to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:25 +0000 UTC Normal Pod my-jaeger-b7b4b4b99-54rs6 AddedInterface Add eth0 [10.131.0.33/23] from ovn-kubernetes multus logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:25 +0000 UTC Normal Pod my-jaeger-b7b4b4b99-54rs6.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:25 +0000 UTC Normal ReplicaSet.apps my-jaeger-b7b4b4b99 SuccessfulCreate Created pod: my-jaeger-b7b4b4b99-54rs6 replicaset-controller logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:25 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-b7b4b4b99 to 1 deployment-controller logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:28 +0000 UTC Normal Pod my-jaeger-b7b4b4b99-54rs6.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" in 2.377s (2.377s including waiting) kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:28 +0000 UTC Normal Pod my-jaeger-b7b4b4b99-54rs6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:28 +0000 UTC Normal Pod my-jaeger-b7b4b4b99-54rs6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:28 +0000 UTC Normal Pod my-jaeger-b7b4b4b99-54rs6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:28 +0000 UTC Normal Pod my-jaeger-b7b4b4b99-54rs6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:28 +0000 UTC Normal Pod my-jaeger-b7b4b4b99-54rs6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:32 +0000 UTC Normal Pod my-jaeger-b7b4b4b99-54rs6.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:32 +0000 UTC Normal Pod my-jaeger-b7b4b4b99-54rs6.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:32 +0000 UTC Normal ReplicaSet.apps my-jaeger-b7b4b4b99 SuccessfulDelete Deleted pod: my-jaeger-b7b4b4b99-54rs6 replicaset-controller logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:32 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-b7b4b4b99 to 0 from 1 deployment-controller logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:33 +0000 UTC Normal Pod my-jaeger-57464b4c54-4tqnw Binding Scheduled Successfully assigned kuttl-test-sweeping-mallard/my-jaeger-57464b4c54-4tqnw to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:33 +0000 UTC Normal ReplicaSet.apps my-jaeger-57464b4c54 SuccessfulCreate Created pod: my-jaeger-57464b4c54-4tqnw replicaset-controller logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:33 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-57464b4c54 to 1 deployment-controller logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:34 +0000 UTC Normal Pod my-jaeger-57464b4c54-4tqnw AddedInterface Add eth0 [10.128.2.38/23] from ovn-kubernetes multus logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:34 +0000 UTC Normal Pod my-jaeger-57464b4c54-4tqnw.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:36 +0000 UTC Normal Pod my-jaeger-57464b4c54-4tqnw.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" in 2.643s (2.643s including waiting) kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:37 +0000 UTC Normal Pod my-jaeger-57464b4c54-4tqnw.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:37 +0000 UTC Normal Pod my-jaeger-57464b4c54-4tqnw.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:37 +0000 UTC Normal Pod my-jaeger-57464b4c54-4tqnw.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:37 +0000 UTC Normal Pod my-jaeger-57464b4c54-4tqnw.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:37 +0000 UTC Normal Pod my-jaeger-57464b4c54-4tqnw.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:39 +0000 UTC Normal Pod check-span-ptgz8 Binding Scheduled Successfully assigned kuttl-test-sweeping-mallard/check-span-ptgz8 to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:39 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-ptgz8 job-controller logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:39 +0000 UTC Normal Pod report-span-nhtfv Binding Scheduled Successfully assigned kuttl-test-sweeping-mallard/report-span-nhtfv to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:39 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-nhtfv job-controller logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:40 +0000 UTC Normal Pod check-span-ptgz8 AddedInterface Add eth0 [10.129.2.35/23] from ovn-kubernetes multus logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:40 +0000 UTC Normal Pod check-span-ptgz8.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:40 +0000 UTC Normal Pod check-span-ptgz8.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 367ms (367ms including waiting) kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:40 +0000 UTC Normal Pod report-span-nhtfv AddedInterface Add eth0 [10.131.0.34/23] from ovn-kubernetes multus logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:40 +0000 UTC Normal Pod report-span-nhtfv.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:40 +0000 UTC Normal Pod report-span-nhtfv.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 399ms (399ms including waiting) kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:41 +0000 UTC Normal Pod check-span-ptgz8.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:41 +0000 UTC Normal Pod check-span-ptgz8.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:41 +0000 UTC Normal Pod report-span-nhtfv.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:41 +0000 UTC Normal Pod report-span-nhtfv.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:51 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:55 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestsweepingmallardmyjaeger-1-69dcc797dd SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestsweepingmallardmyjaeger-1-69dccstjjv replicaset-controller logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsweepingmallardmyjaeger-1-69dccstjjv Binding Scheduled Successfully assigned kuttl-test-sweeping-mallard/elasticsearch-cdm-kuttltestsweepingmallardmyjaeger-1-69dccstjjv to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:55 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestsweepingmallardmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestsweepingmallardmyjaeger-1-69dcc797dd to 1 deployment-controller logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsweepingmallardmyjaeger-1-69dccstjjv AddedInterface Add eth0 [10.129.2.36/23] from ovn-kubernetes multus logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsweepingmallardmyjaeger-1-69dccstjjv.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsweepingmallardmyjaeger-1-69dccstjjv.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsweepingmallardmyjaeger-1-69dccstjjv.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsweepingmallardmyjaeger-1-69dccstjjv.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsweepingmallardmyjaeger-1-69dccstjjv.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:17:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsweepingmallardmyjaeger-1-69dccstjjv.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:11 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsweepingmallardmyjaeger-1-69dccstjjv.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:13 +0000 UTC Normal Job.batch report-span Completed Job completed job-controller logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:22 +0000 UTC Normal Pod my-jaeger-57464b4c54-4tqnw.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:22 +0000 UTC Normal Pod my-jaeger-57464b4c54-4tqnw.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:22 +0000 UTC Normal Pod my-jaeger-collector-8694f74447-x7b74 Binding Scheduled Successfully assigned kuttl-test-sweeping-mallard/my-jaeger-collector-8694f74447-x7b74 to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:22 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-8694f74447 SuccessfulCreate Created pod: my-jaeger-collector-8694f74447-x7b74 replicaset-controller logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:22 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-8694f74447 to 1 deployment-controller logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:22 +0000 UTC Normal Pod my-jaeger-query-549cf44cd7-7bwm8 Binding Scheduled Successfully assigned kuttl-test-sweeping-mallard/my-jaeger-query-549cf44cd7-7bwm8 to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:22 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-549cf44cd7 SuccessfulCreate Created pod: my-jaeger-query-549cf44cd7-7bwm8 replicaset-controller logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:22 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-549cf44cd7 to 1 deployment-controller logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:23 +0000 UTC Normal Pod my-jaeger-collector-8694f74447-x7b74 AddedInterface Add eth0 [10.131.0.35/23] from ovn-kubernetes multus logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:23 +0000 UTC Normal Pod my-jaeger-collector-8694f74447-x7b74.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" already present on machine kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:23 +0000 UTC Normal Pod my-jaeger-collector-8694f74447-x7b74.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:23 +0000 UTC Normal Pod my-jaeger-collector-8694f74447-x7b74.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:23 +0000 UTC Normal Pod my-jaeger-query-549cf44cd7-7bwm8 AddedInterface Add eth0 [10.131.0.36/23] from ovn-kubernetes multus logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:23 +0000 UTC Normal Pod my-jaeger-query-549cf44cd7-7bwm8.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:23 +0000 UTC Normal Pod my-jaeger-query-549cf44cd7-7bwm8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:23 +0000 UTC Normal Pod my-jaeger-query-549cf44cd7-7bwm8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:23 +0000 UTC Normal Pod my-jaeger-query-549cf44cd7-7bwm8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:23 +0000 UTC Normal Pod my-jaeger-query-549cf44cd7-7bwm8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:23 +0000 UTC Normal Pod my-jaeger-query-549cf44cd7-7bwm8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:23 +0000 UTC Normal Pod my-jaeger-query-549cf44cd7-7bwm8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:23 +0000 UTC Normal Pod my-jaeger-query-549cf44cd7-7bwm8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | 2024-06-10 07:18:23 +0000 UTC Normal Pod my-jaeger-query-549cf44cd7-7bwm8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:18:34 | es-from-aio-to-production | Deleting namespace: kuttl-test-sweeping-mallard === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (1229.00s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.96s) --- PASS: kuttl/harness/es-multiinstance (117.32s) --- PASS: kuttl/harness/es-simple-prod (5.92s) --- PASS: kuttl/harness/es-rollover-autoprov (249.92s) --- PASS: kuttl/harness/es-increasing-replicas (111.05s) --- FAIL: kuttl/harness/es-index-cleaner-autoprov (659.20s) --- PASS: kuttl/harness/es-from-aio-to-production (79.43s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name elasticsearch --report --output /logs/artifacts/elasticsearch.xml ./artifacts/kuttl-report.xml time="2024-06-10T07:18:42Z" level=debug msg="Setting a new name for the test suites" time="2024-06-10T07:18:42Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-06-10T07:18:42Z" level=debug msg="normalizing test case names" time="2024-06-10T07:18:42Z" level=debug msg="elasticsearch/artifacts -> elasticsearch_artifacts" time="2024-06-10T07:18:42Z" level=debug msg="elasticsearch/es-multiinstance -> elasticsearch_es_multiinstance" time="2024-06-10T07:18:42Z" level=debug msg="elasticsearch/es-simple-prod -> elasticsearch_es_simple_prod" time="2024-06-10T07:18:42Z" level=debug msg="elasticsearch/es-rollover-autoprov -> elasticsearch_es_rollover_autoprov" time="2024-06-10T07:18:42Z" level=debug msg="elasticsearch/es-increasing-replicas -> elasticsearch_es_increasing_replicas" time="2024-06-10T07:18:42Z" level=debug msg="elasticsearch/es-index-cleaner-autoprov -> elasticsearch_es_index_cleaner_autoprov" time="2024-06-10T07:18:42Z" level=debug msg="elasticsearch/es-from-aio-to-production -> elasticsearch_es_from_aio_to_production" +-----------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------+--------+ | elasticsearch_artifacts | passed | | elasticsearch_es_multiinstance | passed | | elasticsearch_es_simple_prod | passed | | elasticsearch_es_rollover_autoprov | passed | | elasticsearch_es_increasing_replicas | passed | | elasticsearch_es_index_cleaner_autoprov | failed | | elasticsearch_es_from_aio_to_production | passed | +-----------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh examples false true + '[' 3 -ne 3 ']' + test_suite_name=examples + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/examples.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-examples make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ VERTX_IMG=jaegertracing/vertx-create-span:operator-e2e-tests \ ./tests/e2e/examples/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-07-024651 True False 32m Cluster version is 4.16.0-0.nightly-2024-06-07-024651' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-07-024651 True False 32m Cluster version is 4.16.0-0.nightly-2024-06-07-024651' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/examples/render.sh ++ export SUITE_DIR=./tests/e2e/examples ++ SUITE_DIR=./tests/e2e/examples ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/examples ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test examples-agent-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-agent-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-with-priority-class\e[0m' Rendering files for test examples-agent-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + mkdir -p examples-agent-with-priority-class + cd examples-agent-with-priority-class + example_name=agent-with-priority-class + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + '[' true '!=' true ']' + render_install_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + install_file=./02-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/agent-with-priority-class.yaml -o ./02-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./02-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./02-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./02-install.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + JAEGER_NAME=agent-as-daemonset + local jaeger_strategy ++ get_jaeger_strategy ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./02-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./02-install.yaml ++ strategy=DaemonSet ++ '[' DaemonSet = null ']' ++ echo DaemonSet ++ return 0 + jaeger_strategy=DaemonSet + '[' DaemonSet = DaemonSet ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./02-assert.yaml + render_smoke_test_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + jaeger_name=agent-as-daemonset + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test agent-as-daemonset true 02 + '[' 3 -ne 3 ']' + jaeger=agent-as-daemonset + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-all-in-one-with-options + '[' 1 -ne 1 ']' + test_name=examples-all-in-one-with-options + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-all-in-one-with-options' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-all-in-one-with-options\e[0m' Rendering files for test examples-all-in-one-with-options + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-with-priority-class + '[' examples-agent-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-all-in-one-with-options + cd examples-all-in-one-with-options + example_name=all-in-one-with-options + render_install_example all-in-one-with-options 00 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/all-in-one-with-options.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + JAEGER_NAME=my-jaeger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.metadata.name="my-jaeger"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i 'del(.spec.allInOne.image)' ./00-install.yaml + render_smoke_test_example all-in-one-with-options 01 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + jaeger_name=my-jaeger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + sed -i s~my-jaeger-query:443~my-jaeger-query:443/jaeger~gi ./01-smoke-test.yaml + start_test examples-business-application-injected-sidecar + '[' 1 -ne 1 ']' + test_name=examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-business-application-injected-sidecar' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-business-application-injected-sidecar\e[0m' Rendering files for test examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-all-in-one-with-options + '[' examples-all-in-one-with-options '!=' _build ']' + cd .. + mkdir -p examples-business-application-injected-sidecar + cd examples-business-application-injected-sidecar + example_name=simplest + cp /tmp/jaeger-tests/examples/business-application-injected-sidecar.yaml ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].image=strenv(VERTX_IMG)' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.path="/"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.port=8080' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.initialDelaySeconds=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.failureThreshold=3' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.periodSeconds=10' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.successThreshold=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.timeoutSeconds=1' ./00-install.yaml + render_install_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example simplest 02 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 02 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-collector-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-collector-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-collector-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-collector-with-priority-class\e[0m' Rendering files for test examples-collector-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-business-application-injected-sidecar + '[' examples-business-application-injected-sidecar '!=' _build ']' + cd .. + mkdir -p examples-collector-with-priority-class + cd examples-collector-with-priority-class + example_name=collector-with-priority-class + render_install_example collector-with-priority-class 00 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/collector-with-priority-class.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + JAEGER_NAME=collector-with-high-priority + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example collector-with-priority-class 01 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + jaeger_name=collector-with-high-priority + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test collector-with-high-priority true 01 + '[' 3 -ne 3 ']' + jaeger=collector-with-high-priority + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + export JAEGER_NAME=collector-with-high-priority + JAEGER_NAME=collector-with-high-priority + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-service-types + '[' 1 -ne 1 ']' + test_name=examples-service-types + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-service-types' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-service-types\e[0m' Rendering files for test examples-service-types + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-collector-with-priority-class + '[' examples-collector-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-service-types + cd examples-service-types + example_name=service-types + render_install_example service-types 00 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/service-types.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + JAEGER_NAME=service-types + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example service-types 01 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/service-types.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/service-types.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/service-types.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/service-types.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + jaeger_name=service-types + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test service-types true 01 + '[' 3 -ne 3 ']' + jaeger=service-types + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + export JAEGER_NAME=service-types + JAEGER_NAME=service-types + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod + '[' 1 -ne 1 ']' + test_name=examples-simple-prod + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod\e[0m' Rendering files for test examples-simple-prod + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-service-types + '[' examples-service-types '!=' _build ']' + cd .. + mkdir -p examples-simple-prod + cd examples-simple-prod + example_name=simple-prod + render_install_example simple-prod 01 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod 02 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod-with-volumes + '[' 1 -ne 1 ']' + test_name=examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod-with-volumes' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod-with-volumes\e[0m' Rendering files for test examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod + '[' examples-simple-prod '!=' _build ']' + cd .. + mkdir -p examples-simple-prod-with-volumes + cd examples-simple-prod-with-volumes + example_name=simple-prod-with-volumes + render_install_example simple-prod-with-volumes 01 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod-with-volumes 02 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + /tmp/jaeger-tests/bin/gomplate -f ./03-check-volume.yaml.template -o 03-check-volume.yaml + start_test examples-simplest + '[' 1 -ne 1 ']' + test_name=examples-simplest + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simplest' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simplest\e[0m' Rendering files for test examples-simplest + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod-with-volumes + '[' examples-simple-prod-with-volumes '!=' _build ']' + cd .. + mkdir -p examples-simplest + cd examples-simplest + example_name=simplest + render_install_example simplest 00 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 01 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger + '[' 1 -ne 1 ']' + test_name=examples-with-badger + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger\e[0m' Rendering files for test examples-with-badger + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simplest + '[' examples-simplest '!=' _build ']' + cd .. + mkdir -p examples-with-badger + cd examples-with-badger + example_name=with-badger + render_install_example with-badger 00 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + JAEGER_NAME=with-badger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger 01 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + jaeger_name=with-badger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + export JAEGER_NAME=with-badger + JAEGER_NAME=with-badger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger-and-volume + '[' 1 -ne 1 ']' + test_name=examples-with-badger-and-volume + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger-and-volume' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger-and-volume\e[0m' Rendering files for test examples-with-badger-and-volume + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger + '[' examples-with-badger '!=' _build ']' + cd .. + mkdir -p examples-with-badger-and-volume + cd examples-with-badger-and-volume + example_name=with-badger-and-volume + render_install_example with-badger-and-volume 00 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger-and-volume.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + JAEGER_NAME=with-badger-and-volume + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger-and-volume 01 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + jaeger_name=with-badger-and-volume + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger-and-volume true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger-and-volume + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + export JAEGER_NAME=with-badger-and-volume + JAEGER_NAME=with-badger-and-volume + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-cassandra + '[' 1 -ne 1 ']' + test_name=examples-with-cassandra + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-cassandra' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-cassandra\e[0m' Rendering files for test examples-with-cassandra + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger-and-volume + '[' examples-with-badger-and-volume '!=' _build ']' + cd .. + mkdir -p examples-with-cassandra + cd examples-with-cassandra + example_name=with-cassandra + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-cassandra 01 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-cassandra.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + JAEGER_NAME=with-cassandra + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-cassandra 02 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-cassandra.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-cassandra.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + jaeger_name=with-cassandra + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-cassandra true 02 + '[' 3 -ne 3 ']' + jaeger=with-cassandra + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + export JAEGER_NAME=with-cassandra + JAEGER_NAME=with-cassandra + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-sampling + '[' 1 -ne 1 ']' + test_name=examples-with-sampling + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-sampling' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-sampling\e[0m' Rendering files for test examples-with-sampling + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-cassandra + '[' examples-with-cassandra '!=' _build ']' + cd .. + mkdir -p examples-with-sampling + cd examples-with-sampling + export example_name=with-sampling + example_name=with-sampling + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-sampling 01 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-sampling.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + JAEGER_NAME=with-sampling + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-sampling 02 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-sampling.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-sampling.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + jaeger_name=with-sampling + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-sampling true 02 + '[' 3 -ne 3 ']' + jaeger=with-sampling + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + export JAEGER_NAME=with-sampling + JAEGER_NAME=with-sampling + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-agent-as-daemonset + '[' 1 -ne 1 ']' + test_name=examples-agent-as-daemonset + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-as-daemonset' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-as-daemonset\e[0m' Rendering files for test examples-agent-as-daemonset + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-sampling + '[' examples-with-sampling '!=' _build ']' + cd .. + mkdir -p examples-agent-as-daemonset + cd examples-agent-as-daemonset + '[' true = true ']' + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/agent-as-daemonset.yaml -o 02-install.yaml + '[' true = true ']' + start_test examples-openshift-with-htpasswd + '[' 1 -ne 1 ']' + test_name=examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-openshift-with-htpasswd' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-openshift-with-htpasswd\e[0m' Rendering files for test examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-as-daemonset + '[' examples-agent-as-daemonset '!=' _build ']' + cd .. + mkdir -p examples-openshift-with-htpasswd + cd examples-openshift-with-htpasswd + export JAEGER_NAME=with-htpasswd + JAEGER_NAME=with-htpasswd + export JAEGER_USERNAME=awesomeuser + JAEGER_USERNAME=awesomeuser + export JAEGER_PASSWORD=awesomepassword + JAEGER_PASSWORD=awesomepassword + export 'JAEGER_USER_PASSWORD_HASH=awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' + JAEGER_USER_PASSWORD_HASH='awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ echo 'awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ base64 + SECRET=YXdlc29tZXVzZXI6e1NIQX11VWRxUFZVeXFOQm1FUlUwUXhqM0tGYVpuanc9Cg== + /tmp/jaeger-tests/bin/gomplate -f ./00-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/with-htpasswd.yaml -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + export 'GET_URL_COMMAND=kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + export 'URL=https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + INSECURE=true + JAEGER_USERNAME= + JAEGER_PASSWORD= + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-unsecured.yaml + JAEGER_USERNAME=wronguser + JAEGER_PASSWORD=wrongpassword + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-check-unauthorized.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./04-check-authorized.yaml + skip_test examples-agent-as-daemonset 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-agent-as-daemonset + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-openshift-with-htpasswd + '[' examples-openshift-with-htpasswd '!=' _build ']' + cd .. + rm -rf examples-agent-as-daemonset + warning 'examples-agent-as-daemonset: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-agent-as-daemonset: This test is flaky in Prow CI\e[0m' WAR: examples-agent-as-daemonset: This test is flaky in Prow CI + skip_test examples-with-badger-and-volume 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-with-badger-and-volume + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-with-badger-and-volume + warning 'examples-with-badger-and-volume: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-with-badger-and-volume: This test is flaky in Prow CI\e[0m' WAR: examples-with-badger-and-volume: This test is flaky in Prow CI + skip_test examples-collector-with-priority-class 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-collector-with-priority-class + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-collector-with-priority-class + warning 'examples-collector-with-priority-class: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-collector-with-priority-class: This test is flaky in Prow CI\e[0m' WAR: examples-collector-with-priority-class: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running examples E2E tests' Running examples E2E tests + cd tests/e2e/examples/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3357233117 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 12 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/examples-agent-with-priority-class === PAUSE kuttl/harness/examples-agent-with-priority-class === RUN kuttl/harness/examples-all-in-one-with-options === PAUSE kuttl/harness/examples-all-in-one-with-options === RUN kuttl/harness/examples-business-application-injected-sidecar === PAUSE kuttl/harness/examples-business-application-injected-sidecar === RUN kuttl/harness/examples-openshift-with-htpasswd === PAUSE kuttl/harness/examples-openshift-with-htpasswd === RUN kuttl/harness/examples-service-types === PAUSE kuttl/harness/examples-service-types === RUN kuttl/harness/examples-simple-prod === PAUSE kuttl/harness/examples-simple-prod === RUN kuttl/harness/examples-simple-prod-with-volumes === PAUSE kuttl/harness/examples-simple-prod-with-volumes === RUN kuttl/harness/examples-simplest === PAUSE kuttl/harness/examples-simplest === RUN kuttl/harness/examples-with-badger === PAUSE kuttl/harness/examples-with-badger === RUN kuttl/harness/examples-with-cassandra === PAUSE kuttl/harness/examples-with-cassandra === RUN kuttl/harness/examples-with-sampling === PAUSE kuttl/harness/examples-with-sampling === CONT kuttl/harness/artifacts logger.go:42: 07:19:13 | artifacts | Creating namespace: kuttl-test-legal-alpaca logger.go:42: 07:19:13 | artifacts | artifacts events from ns kuttl-test-legal-alpaca: logger.go:42: 07:19:13 | artifacts | Deleting namespace: kuttl-test-legal-alpaca === CONT kuttl/harness/examples-simple-prod logger.go:42: 07:19:19 | examples-simple-prod | Creating namespace: kuttl-test-closing-seagull logger.go:42: 07:19:19 | examples-simple-prod/1-install | starting test step 1-install logger.go:42: 07:19:19 | examples-simple-prod/1-install | Jaeger:kuttl-test-closing-seagull/simple-prod created logger.go:42: 07:19:54 | examples-simple-prod/1-install | test step completed 1-install logger.go:42: 07:19:54 | examples-simple-prod/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:19:54 | examples-simple-prod/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:19:56 | examples-simple-prod/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:20:02 | examples-simple-prod/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:20:03 | examples-simple-prod/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:20:03 | examples-simple-prod/2-smoke-test | job.batch/report-span created logger.go:42: 07:20:03 | examples-simple-prod/2-smoke-test | job.batch/check-span created logger.go:42: 07:20:15 | examples-simple-prod/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:20:16 | examples-simple-prod | examples-simple-prod events from ns kuttl-test-closing-seagull: logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:25 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestclosingseagullsimpleprod-1-84b8bfb668 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestclosingseagullsimpleprod-1-84b8x68vd replicaset-controller logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclosingseagullsimpleprod-1-84b8x68vd Binding Scheduled Successfully assigned kuttl-test-closing-seagull/elasticsearch-cdm-kuttltestclosingseagullsimpleprod-1-84b8x68vd to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclosingseagullsimpleprod-1-84b8x68vd AddedInterface Add eth0 [10.131.0.37/23] from ovn-kubernetes multus logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclosingseagullsimpleprod-1-84b8x68vd.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclosingseagullsimpleprod-1-84b8x68vd.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclosingseagullsimpleprod-1-84b8x68vd.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclosingseagullsimpleprod-1-84b8x68vd.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:25 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestclosingseagullsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestclosingseagullsimpleprod-1-84b8bfb668 to 1 deployment-controller logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclosingseagullsimpleprod-1-84b8x68vd.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclosingseagullsimpleprod-1-84b8x68vd.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:35 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestclosingseagullsimpleprod-1-84b8x68vd.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:40 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestclosingseagullsimpleprod-1-84b8x68vd.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:52 +0000 UTC Normal Pod simple-prod-collector-75f777668b-m8mf9 Binding Scheduled Successfully assigned kuttl-test-closing-seagull/simple-prod-collector-75f777668b-m8mf9 to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:52 +0000 UTC Warning Pod simple-prod-collector-75f777668b-m8mf9 FailedMount MountVolume.SetUp failed for volume "simple-prod-collector-tls-config-volume" : secret "simple-prod-collector-headless-tls" not found kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:52 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-75f777668b SuccessfulCreate Created pod: simple-prod-collector-75f777668b-m8mf9 replicaset-controller logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:52 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-75f777668b to 1 deployment-controller logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:52 +0000 UTC Normal Pod simple-prod-query-866ddf9fd7-7tvqf Binding Scheduled Successfully assigned kuttl-test-closing-seagull/simple-prod-query-866ddf9fd7-7tvqf to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:52 +0000 UTC Normal Pod simple-prod-query-866ddf9fd7-7tvqf AddedInterface Add eth0 [10.128.2.39/23] from ovn-kubernetes multus logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:52 +0000 UTC Normal Pod simple-prod-query-866ddf9fd7-7tvqf.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:52 +0000 UTC Normal Pod simple-prod-query-866ddf9fd7-7tvqf.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:52 +0000 UTC Normal ReplicaSet.apps simple-prod-query-866ddf9fd7 SuccessfulCreate Created pod: simple-prod-query-866ddf9fd7-7tvqf replicaset-controller logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:52 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-866ddf9fd7 to 1 deployment-controller logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:53 +0000 UTC Normal Pod simple-prod-collector-75f777668b-m8mf9 AddedInterface Add eth0 [10.129.2.37/23] from ovn-kubernetes multus logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:53 +0000 UTC Normal Pod simple-prod-collector-75f777668b-m8mf9.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" already present on machine kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:53 +0000 UTC Normal Pod simple-prod-collector-75f777668b-m8mf9.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:53 +0000 UTC Normal Pod simple-prod-collector-75f777668b-m8mf9.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:53 +0000 UTC Normal Pod simple-prod-query-866ddf9fd7-7tvqf.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:53 +0000 UTC Normal Pod simple-prod-query-866ddf9fd7-7tvqf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:53 +0000 UTC Normal Pod simple-prod-query-866ddf9fd7-7tvqf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:53 +0000 UTC Normal Pod simple-prod-query-866ddf9fd7-7tvqf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:53 +0000 UTC Normal Pod simple-prod-query-866ddf9fd7-7tvqf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:53 +0000 UTC Normal Pod simple-prod-query-866ddf9fd7-7tvqf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:53 +0000 UTC Normal Pod simple-prod-query-866ddf9fd7-7tvqf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:58 +0000 UTC Normal Pod simple-prod-query-866ddf9fd7-7tvqf.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:58 +0000 UTC Normal Pod simple-prod-query-866ddf9fd7-7tvqf.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:58 +0000 UTC Normal Pod simple-prod-query-866ddf9fd7-7tvqf.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:58 +0000 UTC Normal ReplicaSet.apps simple-prod-query-866ddf9fd7 SuccessfulDelete Deleted pod: simple-prod-query-866ddf9fd7-7tvqf replicaset-controller logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:58 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-866ddf9fd7 to 0 from 1 deployment-controller logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:59 +0000 UTC Normal Pod simple-prod-query-7b4b464b96-9zzwm Binding Scheduled Successfully assigned kuttl-test-closing-seagull/simple-prod-query-7b4b464b96-9zzwm to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:59 +0000 UTC Normal Pod simple-prod-query-7b4b464b96-9zzwm AddedInterface Add eth0 [10.128.2.40/23] from ovn-kubernetes multus logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:59 +0000 UTC Normal Pod simple-prod-query-7b4b464b96-9zzwm.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:59 +0000 UTC Normal ReplicaSet.apps simple-prod-query-7b4b464b96 SuccessfulCreate Created pod: simple-prod-query-7b4b464b96-9zzwm replicaset-controller logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:19:59 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-7b4b464b96 to 1 deployment-controller logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:00 +0000 UTC Normal Pod simple-prod-query-7b4b464b96-9zzwm.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:00 +0000 UTC Normal Pod simple-prod-query-7b4b464b96-9zzwm.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:00 +0000 UTC Normal Pod simple-prod-query-7b4b464b96-9zzwm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:00 +0000 UTC Normal Pod simple-prod-query-7b4b464b96-9zzwm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:00 +0000 UTC Normal Pod simple-prod-query-7b4b464b96-9zzwm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:00 +0000 UTC Normal Pod simple-prod-query-7b4b464b96-9zzwm.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:00 +0000 UTC Normal Pod simple-prod-query-7b4b464b96-9zzwm.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:00 +0000 UTC Normal Pod simple-prod-query-7b4b464b96-9zzwm.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:03 +0000 UTC Normal Pod check-span-wqkfl Binding Scheduled Successfully assigned kuttl-test-closing-seagull/check-span-wqkfl to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:03 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-wqkfl job-controller logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:03 +0000 UTC Normal Pod report-span-xq9ls Binding Scheduled Successfully assigned kuttl-test-closing-seagull/report-span-xq9ls to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:03 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-xq9ls job-controller logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:04 +0000 UTC Normal Pod check-span-wqkfl AddedInterface Add eth0 [10.129.2.39/23] from ovn-kubernetes multus logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:04 +0000 UTC Normal Pod check-span-wqkfl.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:04 +0000 UTC Normal Pod check-span-wqkfl.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 335ms (335ms including waiting) kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:04 +0000 UTC Normal Pod check-span-wqkfl.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:04 +0000 UTC Normal Pod check-span-wqkfl.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:04 +0000 UTC Normal Pod report-span-xq9ls AddedInterface Add eth0 [10.129.2.38/23] from ovn-kubernetes multus logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:04 +0000 UTC Normal Pod report-span-xq9ls.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:04 +0000 UTC Normal Pod report-span-xq9ls.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 372ms (372ms including waiting) kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:04 +0000 UTC Normal Pod report-span-xq9ls.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:04 +0000 UTC Normal Pod report-span-xq9ls.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:07 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:07 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:07 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:20:16 | examples-simple-prod | 2024-06-10 07:20:15 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:20:16 | examples-simple-prod | Deleting namespace: kuttl-test-closing-seagull === CONT kuttl/harness/examples-with-sampling logger.go:42: 07:20:23 | examples-with-sampling | Creating namespace: kuttl-test-integral-ringtail logger.go:42: 07:20:23 | examples-with-sampling/0-install | starting test step 0-install logger.go:42: 07:20:23 | examples-with-sampling/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:20:23 | examples-with-sampling/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:20:23 | examples-with-sampling/0-install | >>>> Creating namespace kuttl-test-integral-ringtail logger.go:42: 07:20:23 | examples-with-sampling/0-install | kubectl create namespace kuttl-test-integral-ringtail 2>&1 | grep -v "already exists" || true logger.go:42: 07:20:23 | examples-with-sampling/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-integral-ringtail 2>&1 | grep -v "already exists" || true logger.go:42: 07:20:23 | examples-with-sampling/0-install | service/cassandra created logger.go:42: 07:20:23 | examples-with-sampling/0-install | statefulset.apps/cassandra created logger.go:42: 07:20:23 | examples-with-sampling/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:20:52 | examples-with-sampling/0-install | test step completed 0-install logger.go:42: 07:20:52 | examples-with-sampling/1-install | starting test step 1-install logger.go:42: 07:20:52 | examples-with-sampling/1-install | Jaeger:kuttl-test-integral-ringtail/with-sampling created logger.go:42: 07:20:57 | examples-with-sampling/1-install | test step completed 1-install logger.go:42: 07:20:57 | examples-with-sampling/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:20:57 | examples-with-sampling/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-sampling /dev/null] logger.go:42: 07:20:59 | examples-with-sampling/2-smoke-test | Warning: resource jaegers/with-sampling is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:21:06 | examples-with-sampling/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:21:06 | examples-with-sampling/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:21:07 | examples-with-sampling/2-smoke-test | job.batch/report-span created logger.go:42: 07:21:07 | examples-with-sampling/2-smoke-test | job.batch/check-span created logger.go:42: 07:21:19 | examples-with-sampling/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:21:19 | examples-with-sampling/3- | starting test step 3- logger.go:42: 07:21:19 | examples-with-sampling/3- | test step completed 3- logger.go:42: 07:21:19 | examples-with-sampling | examples-with-sampling events from ns kuttl-test-integral-ringtail: logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:23 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-integral-ringtail/cassandra-0 to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:23 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:24 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.129.2.40/23] from ovn-kubernetes multus logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:24 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:28 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 4.237s (4.237s including waiting) kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:28 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:28 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:29 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-integral-ringtail/cassandra-1 to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:29 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:30 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:30 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.131.0.38/23] from ovn-kubernetes multus logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:30 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:32 +0000 UTC Warning Pod cassandra-0.spec.containers{cassandra} BackOff Back-off restarting failed container cassandra in pod cassandra-0_kuttl-test-integral-ringtail(cddcf4f7-2bb0-4126-b809-b84e7fe93079) kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:34 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 4.457s (4.457s including waiting) kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:34 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:34 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:36 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:38 +0000 UTC Warning Pod cassandra-1.spec.containers{cassandra} BackOff Back-off restarting failed container cassandra in pod cassandra-1_kuttl-test-integral-ringtail(9f8dc600-61c5-485b-be22-a9d5b657c29a) kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:56 +0000 UTC Normal Pod with-sampling-587695cbd6-6fbjt Binding Scheduled Successfully assigned kuttl-test-integral-ringtail/with-sampling-587695cbd6-6fbjt to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:56 +0000 UTC Normal Pod with-sampling-587695cbd6-6fbjt AddedInterface Add eth0 [10.128.2.41/23] from ovn-kubernetes multus logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:56 +0000 UTC Normal Pod with-sampling-587695cbd6-6fbjt.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:56 +0000 UTC Normal Pod with-sampling-587695cbd6-6fbjt.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:56 +0000 UTC Normal Pod with-sampling-587695cbd6-6fbjt.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:56 +0000 UTC Normal Pod with-sampling-587695cbd6-6fbjt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:56 +0000 UTC Normal Pod with-sampling-587695cbd6-6fbjt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:56 +0000 UTC Normal Pod with-sampling-587695cbd6-6fbjt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:56 +0000 UTC Normal ReplicaSet.apps with-sampling-587695cbd6 SuccessfulCreate Created pod: with-sampling-587695cbd6-6fbjt replicaset-controller logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:20:56 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-587695cbd6 to 1 deployment-controller logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:03 +0000 UTC Normal Pod with-sampling-587695cbd6-6fbjt.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:03 +0000 UTC Normal Pod with-sampling-587695cbd6-6fbjt.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:03 +0000 UTC Normal ReplicaSet.apps with-sampling-587695cbd6 SuccessfulDelete Deleted pod: with-sampling-587695cbd6-6fbjt replicaset-controller logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:03 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled down replica set with-sampling-587695cbd6 to 0 from 1 deployment-controller logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:04 +0000 UTC Normal Pod with-sampling-744f5f4ccd-d6gc5 Binding Scheduled Successfully assigned kuttl-test-integral-ringtail/with-sampling-744f5f4ccd-d6gc5 to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:04 +0000 UTC Normal ReplicaSet.apps with-sampling-744f5f4ccd SuccessfulCreate Created pod: with-sampling-744f5f4ccd-d6gc5 replicaset-controller logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:04 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-744f5f4ccd to 1 deployment-controller logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:05 +0000 UTC Normal Pod with-sampling-744f5f4ccd-d6gc5 AddedInterface Add eth0 [10.128.2.42/23] from ovn-kubernetes multus logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:05 +0000 UTC Normal Pod with-sampling-744f5f4ccd-d6gc5.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:05 +0000 UTC Normal Pod with-sampling-744f5f4ccd-d6gc5.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:05 +0000 UTC Normal Pod with-sampling-744f5f4ccd-d6gc5.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:05 +0000 UTC Normal Pod with-sampling-744f5f4ccd-d6gc5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:05 +0000 UTC Normal Pod with-sampling-744f5f4ccd-d6gc5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:05 +0000 UTC Normal Pod with-sampling-744f5f4ccd-d6gc5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:07 +0000 UTC Normal Pod check-span-tkz6g Binding Scheduled Successfully assigned kuttl-test-integral-ringtail/check-span-tkz6g to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:07 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-tkz6g job-controller logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:07 +0000 UTC Normal Pod report-span-s92n2 Binding Scheduled Successfully assigned kuttl-test-integral-ringtail/report-span-s92n2 to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:07 +0000 UTC Normal Pod report-span-s92n2 AddedInterface Add eth0 [10.129.2.41/23] from ovn-kubernetes multus logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:07 +0000 UTC Normal Pod report-span-s92n2.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:07 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-s92n2 job-controller logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:08 +0000 UTC Normal Pod check-span-tkz6g AddedInterface Add eth0 [10.131.0.39/23] from ovn-kubernetes multus logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:08 +0000 UTC Normal Pod check-span-tkz6g.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:08 +0000 UTC Normal Pod check-span-tkz6g.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 325ms (325ms including waiting) kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:08 +0000 UTC Normal Pod check-span-tkz6g.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:08 +0000 UTC Normal Pod check-span-tkz6g.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:08 +0000 UTC Normal Pod report-span-s92n2.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 307ms (307ms including waiting) kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:08 +0000 UTC Normal Pod report-span-s92n2.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:08 +0000 UTC Normal Pod report-span-s92n2.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:21:19 | examples-with-sampling | 2024-06-10 07:21:19 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:21:19 | examples-with-sampling | Deleting namespace: kuttl-test-integral-ringtail === CONT kuttl/harness/examples-with-cassandra logger.go:42: 07:26:08 | examples-with-cassandra | Creating namespace: kuttl-test-whole-starfish logger.go:42: 07:26:08 | examples-with-cassandra/0-install | starting test step 0-install logger.go:42: 07:26:08 | examples-with-cassandra/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:26:08 | examples-with-cassandra/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:26:08 | examples-with-cassandra/0-install | >>>> Creating namespace kuttl-test-whole-starfish logger.go:42: 07:26:08 | examples-with-cassandra/0-install | kubectl create namespace kuttl-test-whole-starfish 2>&1 | grep -v "already exists" || true logger.go:42: 07:26:09 | examples-with-cassandra/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-whole-starfish 2>&1 | grep -v "already exists" || true logger.go:42: 07:26:09 | examples-with-cassandra/0-install | service/cassandra created logger.go:42: 07:26:09 | examples-with-cassandra/0-install | statefulset.apps/cassandra created logger.go:42: 07:26:09 | examples-with-cassandra/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:26:17 | examples-with-cassandra/0-install | test step completed 0-install logger.go:42: 07:26:17 | examples-with-cassandra/1-install | starting test step 1-install logger.go:42: 07:26:17 | examples-with-cassandra/1-install | Jaeger:kuttl-test-whole-starfish/with-cassandra created logger.go:42: 07:28:00 | examples-with-cassandra/1-install | test step completed 1-install logger.go:42: 07:28:00 | examples-with-cassandra/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:28:00 | examples-with-cassandra/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-cassandra /dev/null] logger.go:42: 07:28:02 | examples-with-cassandra/2-smoke-test | Warning: resource jaegers/with-cassandra is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:28:08 | examples-with-cassandra/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:28:09 | examples-with-cassandra/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:28:09 | examples-with-cassandra/2-smoke-test | job.batch/report-span created logger.go:42: 07:28:09 | examples-with-cassandra/2-smoke-test | job.batch/check-span created logger.go:42: 07:28:21 | examples-with-cassandra/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:28:22 | examples-with-cassandra | examples-with-cassandra events from ns kuttl-test-whole-starfish: logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:09 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-whole-starfish/cassandra-0 to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:09 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:10 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.128.2.43/23] from ovn-kubernetes multus logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:10 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:14 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 4.074s (4.074s including waiting) kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:14 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:14 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:15 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-whole-starfish/cassandra-1 to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:15 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.131.0.40/23] from ovn-kubernetes multus logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:15 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:15 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:15 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:15 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:16 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:18 +0000 UTC Warning Pod cassandra-0.spec.containers{cassandra} BackOff Back-off restarting failed container cassandra in pod cassandra-0_kuttl-test-whole-starfish(fb3cbc7b-54ba-44d8-8eab-9e54ef986634) kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:19 +0000 UTC Warning Pod cassandra-1.spec.containers{cassandra} BackOff Back-off restarting failed container cassandra in pod cassandra-1_kuttl-test-whole-starfish(f88aa5fb-966d-405b-a93a-e1866537a1ee) kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:21 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-89x89 Binding Scheduled Successfully assigned kuttl-test-whole-starfish/with-cassandra-cassandra-schema-job-89x89 to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:21 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-89x89 AddedInterface Add eth0 [10.129.2.42/23] from ovn-kubernetes multus logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:21 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-89x89.spec.containers{with-cassandra-cassandra-schema-job} Pulling Pulling image "jaegertracing/jaeger-cassandra-schema:1.57.0" kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:21 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job SuccessfulCreate Created pod: with-cassandra-cassandra-schema-job-89x89 job-controller logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:27 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-89x89.spec.containers{with-cassandra-cassandra-schema-job} Pulled Successfully pulled image "jaegertracing/jaeger-cassandra-schema:1.57.0" in 5.879s (5.879s including waiting) kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:27 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-89x89.spec.containers{with-cassandra-cassandra-schema-job} Created Created container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:26:27 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-89x89.spec.containers{with-cassandra-cassandra-schema-job} Started Started container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:27:54 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job Completed Job completed job-controller logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:27:55 +0000 UTC Normal Pod with-cassandra-6954b998bb-9n648 Binding Scheduled Successfully assigned kuttl-test-whole-starfish/with-cassandra-6954b998bb-9n648 to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:27:55 +0000 UTC Normal ReplicaSet.apps with-cassandra-6954b998bb SuccessfulCreate Created pod: with-cassandra-6954b998bb-9n648 replicaset-controller logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:27:55 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-6954b998bb to 1 deployment-controller logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:27:56 +0000 UTC Normal Pod with-cassandra-6954b998bb-9n648 AddedInterface Add eth0 [10.129.2.43/23] from ovn-kubernetes multus logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:27:56 +0000 UTC Normal Pod with-cassandra-6954b998bb-9n648.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:27:58 +0000 UTC Normal Pod with-cassandra-6954b998bb-9n648.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" in 2.684s (2.684s including waiting) kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:27:58 +0000 UTC Normal Pod with-cassandra-6954b998bb-9n648.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:27:58 +0000 UTC Normal Pod with-cassandra-6954b998bb-9n648.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:27:58 +0000 UTC Normal Pod with-cassandra-6954b998bb-9n648.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:27:58 +0000 UTC Normal Pod with-cassandra-6954b998bb-9n648.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:27:58 +0000 UTC Normal Pod with-cassandra-6954b998bb-9n648.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:03 +0000 UTC Normal Pod with-cassandra-6954b998bb-9n648.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:03 +0000 UTC Normal Pod with-cassandra-6954b998bb-9n648.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:03 +0000 UTC Normal ReplicaSet.apps with-cassandra-6954b998bb SuccessfulDelete Deleted pod: with-cassandra-6954b998bb-9n648 replicaset-controller logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:03 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled down replica set with-cassandra-6954b998bb to 0 from 1 deployment-controller logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:04 +0000 UTC Normal Pod with-cassandra-595d84878f-7mgkz Binding Scheduled Successfully assigned kuttl-test-whole-starfish/with-cassandra-595d84878f-7mgkz to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:04 +0000 UTC Normal ReplicaSet.apps with-cassandra-595d84878f SuccessfulCreate Created pod: with-cassandra-595d84878f-7mgkz replicaset-controller logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:04 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-595d84878f to 1 deployment-controller logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:05 +0000 UTC Normal Pod with-cassandra-595d84878f-7mgkz AddedInterface Add eth0 [10.128.2.44/23] from ovn-kubernetes multus logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:05 +0000 UTC Normal Pod with-cassandra-595d84878f-7mgkz.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:05 +0000 UTC Normal Pod with-cassandra-595d84878f-7mgkz.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:05 +0000 UTC Normal Pod with-cassandra-595d84878f-7mgkz.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:05 +0000 UTC Normal Pod with-cassandra-595d84878f-7mgkz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:05 +0000 UTC Normal Pod with-cassandra-595d84878f-7mgkz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:05 +0000 UTC Normal Pod with-cassandra-595d84878f-7mgkz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:09 +0000 UTC Normal Pod check-span-kd87t Binding Scheduled Successfully assigned kuttl-test-whole-starfish/check-span-kd87t to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:09 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-kd87t job-controller logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:09 +0000 UTC Normal Pod report-span-cwck8 Binding Scheduled Successfully assigned kuttl-test-whole-starfish/report-span-cwck8 to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:09 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-cwck8 job-controller logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:10 +0000 UTC Normal Pod check-span-kd87t AddedInterface Add eth0 [10.131.0.41/23] from ovn-kubernetes multus logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:10 +0000 UTC Normal Pod check-span-kd87t.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:10 +0000 UTC Normal Pod check-span-kd87t.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 364ms (364ms including waiting) kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:10 +0000 UTC Normal Pod check-span-kd87t.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:10 +0000 UTC Normal Pod check-span-kd87t.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:10 +0000 UTC Normal Pod report-span-cwck8 AddedInterface Add eth0 [10.129.2.44/23] from ovn-kubernetes multus logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:10 +0000 UTC Normal Pod report-span-cwck8.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:10 +0000 UTC Normal Pod report-span-cwck8.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 353ms (353ms including waiting) kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:10 +0000 UTC Normal Pod report-span-cwck8.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:10 +0000 UTC Normal Pod report-span-cwck8.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:28:22 | examples-with-cassandra | 2024-06-10 07:28:21 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:28:22 | examples-with-cassandra | Deleting namespace: kuttl-test-whole-starfish === CONT kuttl/harness/examples-with-badger logger.go:42: 07:28:39 | examples-with-badger | Creating namespace: kuttl-test-intense-kite logger.go:42: 07:28:39 | examples-with-badger/0-install | starting test step 0-install logger.go:42: 07:28:39 | examples-with-badger/0-install | Jaeger:kuttl-test-intense-kite/with-badger created logger.go:42: 07:28:46 | examples-with-badger/0-install | test step completed 0-install logger.go:42: 07:28:46 | examples-with-badger/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:28:46 | examples-with-badger/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger /dev/null] logger.go:42: 07:28:47 | examples-with-badger/1-smoke-test | Warning: resource jaegers/with-badger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:28:53 | examples-with-badger/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:28:54 | examples-with-badger/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:28:54 | examples-with-badger/1-smoke-test | job.batch/report-span created logger.go:42: 07:28:54 | examples-with-badger/1-smoke-test | job.batch/check-span created logger.go:42: 07:29:07 | examples-with-badger/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:29:07 | examples-with-badger | examples-with-badger events from ns kuttl-test-intense-kite: logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:43 +0000 UTC Normal Pod with-badger-8cf5d4647-m6nfs Binding Scheduled Successfully assigned kuttl-test-intense-kite/with-badger-8cf5d4647-m6nfs to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:43 +0000 UTC Normal ReplicaSet.apps with-badger-8cf5d4647 SuccessfulCreate Created pod: with-badger-8cf5d4647-m6nfs replicaset-controller logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:43 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-8cf5d4647 to 1 deployment-controller logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:44 +0000 UTC Normal Pod with-badger-8cf5d4647-m6nfs AddedInterface Add eth0 [10.129.2.45/23] from ovn-kubernetes multus logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:44 +0000 UTC Normal Pod with-badger-8cf5d4647-m6nfs.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:44 +0000 UTC Normal Pod with-badger-8cf5d4647-m6nfs.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:44 +0000 UTC Normal Pod with-badger-8cf5d4647-m6nfs.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:44 +0000 UTC Normal Pod with-badger-8cf5d4647-m6nfs.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:44 +0000 UTC Normal Pod with-badger-8cf5d4647-m6nfs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:44 +0000 UTC Normal Pod with-badger-8cf5d4647-m6nfs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:50 +0000 UTC Normal Pod with-badger-8cf5d4647-m6nfs.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:50 +0000 UTC Normal Pod with-badger-8cf5d4647-m6nfs.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:50 +0000 UTC Normal ReplicaSet.apps with-badger-8cf5d4647 SuccessfulDelete Deleted pod: with-badger-8cf5d4647-m6nfs replicaset-controller logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:50 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled down replica set with-badger-8cf5d4647 to 0 from 1 deployment-controller logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:51 +0000 UTC Normal Pod with-badger-6dd84d69c4-fl2lc Binding Scheduled Successfully assigned kuttl-test-intense-kite/with-badger-6dd84d69c4-fl2lc to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:51 +0000 UTC Normal ReplicaSet.apps with-badger-6dd84d69c4 SuccessfulCreate Created pod: with-badger-6dd84d69c4-fl2lc replicaset-controller logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:51 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-6dd84d69c4 to 1 deployment-controller logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:52 +0000 UTC Normal Pod with-badger-6dd84d69c4-fl2lc AddedInterface Add eth0 [10.129.2.46/23] from ovn-kubernetes multus logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:52 +0000 UTC Normal Pod with-badger-6dd84d69c4-fl2lc.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:52 +0000 UTC Normal Pod with-badger-6dd84d69c4-fl2lc.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:52 +0000 UTC Normal Pod with-badger-6dd84d69c4-fl2lc.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:52 +0000 UTC Normal Pod with-badger-6dd84d69c4-fl2lc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:52 +0000 UTC Normal Pod with-badger-6dd84d69c4-fl2lc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:52 +0000 UTC Normal Pod with-badger-6dd84d69c4-fl2lc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:54 +0000 UTC Normal Pod check-span-gcgcf Binding Scheduled Successfully assigned kuttl-test-intense-kite/check-span-gcgcf to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:54 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-gcgcf job-controller logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:54 +0000 UTC Normal Pod report-span-6dm2p Binding Scheduled Successfully assigned kuttl-test-intense-kite/report-span-6dm2p to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:54 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-6dm2p job-controller logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:55 +0000 UTC Normal Pod check-span-gcgcf AddedInterface Add eth0 [10.131.0.42/23] from ovn-kubernetes multus logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:55 +0000 UTC Normal Pod check-span-gcgcf.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:55 +0000 UTC Normal Pod check-span-gcgcf.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 333ms (333ms including waiting) kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:55 +0000 UTC Normal Pod report-span-6dm2p AddedInterface Add eth0 [10.128.2.45/23] from ovn-kubernetes multus logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:55 +0000 UTC Normal Pod report-span-6dm2p.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:55 +0000 UTC Normal Pod report-span-6dm2p.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 355ms (355ms including waiting) kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:55 +0000 UTC Normal Pod report-span-6dm2p.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:55 +0000 UTC Normal Pod report-span-6dm2p.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:56 +0000 UTC Normal Pod check-span-gcgcf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:28:56 +0000 UTC Normal Pod check-span-gcgcf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:29:07 | examples-with-badger | 2024-06-10 07:29:06 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:29:07 | examples-with-badger | Deleting namespace: kuttl-test-intense-kite === CONT kuttl/harness/examples-simplest logger.go:42: 07:29:19 | examples-simplest | Creating namespace: kuttl-test-apt-foal logger.go:42: 07:29:19 | examples-simplest/0-install | starting test step 0-install logger.go:42: 07:29:19 | examples-simplest/0-install | Jaeger:kuttl-test-apt-foal/simplest created logger.go:42: 07:29:26 | examples-simplest/0-install | test step completed 0-install logger.go:42: 07:29:26 | examples-simplest/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:29:26 | examples-simplest/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:29:28 | examples-simplest/1-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:29:36 | examples-simplest/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:29:37 | examples-simplest/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:29:37 | examples-simplest/1-smoke-test | job.batch/report-span created logger.go:42: 07:29:37 | examples-simplest/1-smoke-test | job.batch/check-span created logger.go:42: 07:29:48 | examples-simplest/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:29:48 | examples-simplest | examples-simplest events from ns kuttl-test-apt-foal: logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:23 +0000 UTC Normal Pod simplest-66d5c9f74-5kdpg Binding Scheduled Successfully assigned kuttl-test-apt-foal/simplest-66d5c9f74-5kdpg to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:23 +0000 UTC Warning Pod simplest-66d5c9f74-5kdpg FailedMount MountVolume.SetUp failed for volume "simplest-ui-oauth-proxy-tls" : secret "simplest-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:23 +0000 UTC Normal ReplicaSet.apps simplest-66d5c9f74 SuccessfulCreate Created pod: simplest-66d5c9f74-5kdpg replicaset-controller logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:23 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-66d5c9f74 to 1 deployment-controller logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:24 +0000 UTC Normal Pod simplest-66d5c9f74-5kdpg AddedInterface Add eth0 [10.129.2.47/23] from ovn-kubernetes multus logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:24 +0000 UTC Normal Pod simplest-66d5c9f74-5kdpg.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:24 +0000 UTC Normal Pod simplest-66d5c9f74-5kdpg.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:24 +0000 UTC Normal Pod simplest-66d5c9f74-5kdpg.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:24 +0000 UTC Normal Pod simplest-66d5c9f74-5kdpg.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:24 +0000 UTC Normal Pod simplest-66d5c9f74-5kdpg.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:24 +0000 UTC Normal Pod simplest-66d5c9f74-5kdpg.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:33 +0000 UTC Normal Pod simplest-66d5c9f74-5kdpg.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:33 +0000 UTC Normal Pod simplest-66d5c9f74-5kdpg.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:33 +0000 UTC Normal ReplicaSet.apps simplest-66d5c9f74 SuccessfulDelete Deleted pod: simplest-66d5c9f74-5kdpg replicaset-controller logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:33 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-66d5c9f74 to 0 from 1 deployment-controller logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:34 +0000 UTC Normal Pod simplest-7df459bf85-qnbwv Binding Scheduled Successfully assigned kuttl-test-apt-foal/simplest-7df459bf85-qnbwv to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:34 +0000 UTC Normal Pod simplest-7df459bf85-qnbwv AddedInterface Add eth0 [10.131.0.43/23] from ovn-kubernetes multus logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:34 +0000 UTC Normal Pod simplest-7df459bf85-qnbwv.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:34 +0000 UTC Normal Pod simplest-7df459bf85-qnbwv.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:34 +0000 UTC Normal Pod simplest-7df459bf85-qnbwv.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:34 +0000 UTC Normal Pod simplest-7df459bf85-qnbwv.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:34 +0000 UTC Normal Pod simplest-7df459bf85-qnbwv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:34 +0000 UTC Normal Pod simplest-7df459bf85-qnbwv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:34 +0000 UTC Normal ReplicaSet.apps simplest-7df459bf85 SuccessfulCreate Created pod: simplest-7df459bf85-qnbwv replicaset-controller logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:34 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-7df459bf85 to 1 deployment-controller logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:37 +0000 UTC Normal Pod check-span-9qsdv Binding Scheduled Successfully assigned kuttl-test-apt-foal/check-span-9qsdv to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:37 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-9qsdv job-controller logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:37 +0000 UTC Normal Pod report-span-z6ftw Binding Scheduled Successfully assigned kuttl-test-apt-foal/report-span-z6ftw to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:37 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-z6ftw job-controller logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:38 +0000 UTC Normal Pod check-span-9qsdv AddedInterface Add eth0 [10.128.2.46/23] from ovn-kubernetes multus logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:38 +0000 UTC Normal Pod check-span-9qsdv.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:38 +0000 UTC Normal Pod check-span-9qsdv.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 357ms (357ms including waiting) kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:38 +0000 UTC Normal Pod check-span-9qsdv.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:38 +0000 UTC Normal Pod check-span-9qsdv.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:38 +0000 UTC Normal Pod report-span-z6ftw AddedInterface Add eth0 [10.129.2.48/23] from ovn-kubernetes multus logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:38 +0000 UTC Normal Pod report-span-z6ftw.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:38 +0000 UTC Normal Pod report-span-z6ftw.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 342ms (342ms including waiting) kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:38 +0000 UTC Normal Pod report-span-z6ftw.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:38 +0000 UTC Normal Pod report-span-z6ftw.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:29:48 | examples-simplest | 2024-06-10 07:29:48 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:29:48 | examples-simplest | Deleting namespace: kuttl-test-apt-foal === CONT kuttl/harness/examples-simple-prod-with-volumes logger.go:42: 07:30:00 | examples-simple-prod-with-volumes | Ignoring 03-check-volume.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:30:00 | examples-simple-prod-with-volumes | Creating namespace: kuttl-test-awake-mayfly logger.go:42: 07:30:00 | examples-simple-prod-with-volumes/1-install | starting test step 1-install logger.go:42: 07:30:00 | examples-simple-prod-with-volumes/1-install | Jaeger:kuttl-test-awake-mayfly/simple-prod created logger.go:42: 07:30:38 | examples-simple-prod-with-volumes/1-install | test step completed 1-install logger.go:42: 07:30:38 | examples-simple-prod-with-volumes/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:30:38 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:30:39 | examples-simple-prod-with-volumes/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:30:46 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:30:47 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:30:47 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/report-span created logger.go:42: 07:30:47 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/check-span created logger.go:42: 07:31:00 | examples-simple-prod-with-volumes/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:31:00 | examples-simple-prod-with-volumes/3-check-volume | starting test step 3-check-volume logger.go:42: 07:31:00 | examples-simple-prod-with-volumes/3-check-volume | running command: [sh -c kubectl exec $(kubectl get pods -n $NAMESPACE -l app=jaeger -l app.kubernetes.io/component=collector -o yaml | /tmp/jaeger-tests/bin/yq e '.items[0].metadata.name') -n $NAMESPACE -- ls /usr/share/elasticsearch/data] logger.go:42: 07:31:00 | examples-simple-prod-with-volumes/3-check-volume | test step completed 3-check-volume logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | examples-simple-prod-with-volumes events from ns kuttl-test-awake-mayfly: logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:07 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestawakemayflysimpleprod-1-7545694998 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestawakemayflysimpleprod-1-7545694n2gx6 replicaset-controller logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:07 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawakemayflysimpleprod-1-7545694n2gx6 Binding Scheduled Successfully assigned kuttl-test-awake-mayfly/elasticsearch-cdm-kuttltestawakemayflysimpleprod-1-7545694n2gx6 to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:07 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestawakemayflysimpleprod-1-7545694n2gx6 FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:07 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestawakemayflysimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestawakemayflysimpleprod-1-7545694998 to 1 deployment-controller logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawakemayflysimpleprod-1-7545694n2gx6 AddedInterface Add eth0 [10.131.0.44/23] from ovn-kubernetes multus logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawakemayflysimpleprod-1-7545694n2gx6.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawakemayflysimpleprod-1-7545694n2gx6.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawakemayflysimpleprod-1-7545694n2gx6.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawakemayflysimpleprod-1-7545694n2gx6.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:09 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawakemayflysimpleprod-1-7545694n2gx6.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:09 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawakemayflysimpleprod-1-7545694n2gx6.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:18 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestawakemayflysimpleprod-1-7545694n2gx6.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:23 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestawakemayflysimpleprod-1-7545694n2gx6.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:34 +0000 UTC Normal Pod simple-prod-collector-65c8865867-xm2hh Binding Scheduled Successfully assigned kuttl-test-awake-mayfly/simple-prod-collector-65c8865867-xm2hh to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:34 +0000 UTC Warning Pod simple-prod-collector-65c8865867-xm2hh FailedMount MountVolume.SetUp failed for volume "simple-prod-collector-tls-config-volume" : secret "simple-prod-collector-headless-tls" not found kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:34 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-65c8865867 SuccessfulCreate Created pod: simple-prod-collector-65c8865867-xm2hh replicaset-controller logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:34 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-65c8865867 to 1 deployment-controller logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:34 +0000 UTC Normal Pod simple-prod-query-75dcd9f6bc-v86hd Binding Scheduled Successfully assigned kuttl-test-awake-mayfly/simple-prod-query-75dcd9f6bc-v86hd to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:34 +0000 UTC Warning Pod simple-prod-query-75dcd9f6bc-v86hd FailedMount MountVolume.SetUp failed for volume "simple-prod-ui-oauth-proxy-tls" : secret "simple-prod-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:34 +0000 UTC Normal ReplicaSet.apps simple-prod-query-75dcd9f6bc SuccessfulCreate Created pod: simple-prod-query-75dcd9f6bc-v86hd replicaset-controller logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:34 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-75dcd9f6bc to 1 deployment-controller logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:35 +0000 UTC Normal Pod simple-prod-collector-65c8865867-xm2hh AddedInterface Add eth0 [10.129.2.49/23] from ovn-kubernetes multus logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:35 +0000 UTC Normal Pod simple-prod-collector-65c8865867-xm2hh.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" already present on machine kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:35 +0000 UTC Normal Pod simple-prod-collector-65c8865867-xm2hh.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:35 +0000 UTC Normal Pod simple-prod-collector-65c8865867-xm2hh.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:35 +0000 UTC Normal Pod simple-prod-query-75dcd9f6bc-v86hd AddedInterface Add eth0 [10.128.2.48/23] from ovn-kubernetes multus logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:35 +0000 UTC Normal Pod simple-prod-query-75dcd9f6bc-v86hd.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:35 +0000 UTC Normal Pod simple-prod-query-75dcd9f6bc-v86hd.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:35 +0000 UTC Normal Pod simple-prod-query-75dcd9f6bc-v86hd.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:35 +0000 UTC Normal Pod simple-prod-query-75dcd9f6bc-v86hd.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:36 +0000 UTC Normal Pod simple-prod-query-75dcd9f6bc-v86hd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:36 +0000 UTC Normal Pod simple-prod-query-75dcd9f6bc-v86hd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:36 +0000 UTC Normal Pod simple-prod-query-75dcd9f6bc-v86hd.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:36 +0000 UTC Normal Pod simple-prod-query-75dcd9f6bc-v86hd.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:36 +0000 UTC Normal Pod simple-prod-query-75dcd9f6bc-v86hd.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:42 +0000 UTC Normal Pod simple-prod-query-75dcd9f6bc-v86hd.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:42 +0000 UTC Normal Pod simple-prod-query-75dcd9f6bc-v86hd.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:42 +0000 UTC Normal Pod simple-prod-query-75dcd9f6bc-v86hd.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:42 +0000 UTC Normal ReplicaSet.apps simple-prod-query-75dcd9f6bc SuccessfulDelete Deleted pod: simple-prod-query-75dcd9f6bc-v86hd replicaset-controller logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:42 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-75dcd9f6bc to 0 from 1 deployment-controller logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:43 +0000 UTC Normal Pod simple-prod-query-5854564749-ttrpq Binding Scheduled Successfully assigned kuttl-test-awake-mayfly/simple-prod-query-5854564749-ttrpq to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:43 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5854564749 SuccessfulCreate Created pod: simple-prod-query-5854564749-ttrpq replicaset-controller logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:43 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-5854564749 to 1 deployment-controller logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:44 +0000 UTC Normal Pod simple-prod-query-5854564749-ttrpq AddedInterface Add eth0 [10.128.2.49/23] from ovn-kubernetes multus logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:44 +0000 UTC Normal Pod simple-prod-query-5854564749-ttrpq.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:44 +0000 UTC Normal Pod simple-prod-query-5854564749-ttrpq.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:44 +0000 UTC Normal Pod simple-prod-query-5854564749-ttrpq.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:44 +0000 UTC Normal Pod simple-prod-query-5854564749-ttrpq.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:44 +0000 UTC Normal Pod simple-prod-query-5854564749-ttrpq.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:44 +0000 UTC Normal Pod simple-prod-query-5854564749-ttrpq.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:44 +0000 UTC Normal Pod simple-prod-query-5854564749-ttrpq.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:45 +0000 UTC Normal Pod simple-prod-query-5854564749-ttrpq.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:45 +0000 UTC Normal Pod simple-prod-query-5854564749-ttrpq.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:47 +0000 UTC Normal Pod check-span-4gcpz Binding Scheduled Successfully assigned kuttl-test-awake-mayfly/check-span-4gcpz to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:47 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-4gcpz job-controller logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:47 +0000 UTC Normal Pod report-span-9nfrm Binding Scheduled Successfully assigned kuttl-test-awake-mayfly/report-span-9nfrm to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:47 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-9nfrm job-controller logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:48 +0000 UTC Normal Pod check-span-4gcpz AddedInterface Add eth0 [10.129.2.51/23] from ovn-kubernetes multus logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:48 +0000 UTC Normal Pod check-span-4gcpz.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:48 +0000 UTC Normal Pod check-span-4gcpz.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 321ms (321ms including waiting) kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:48 +0000 UTC Normal Pod check-span-4gcpz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:48 +0000 UTC Normal Pod check-span-4gcpz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:48 +0000 UTC Normal Pod report-span-9nfrm AddedInterface Add eth0 [10.129.2.50/23] from ovn-kubernetes multus logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:48 +0000 UTC Normal Pod report-span-9nfrm.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:48 +0000 UTC Normal Pod report-span-9nfrm.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 389ms (390ms including waiting) kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:48 +0000 UTC Normal Pod report-span-9nfrm.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:48 +0000 UTC Normal Pod report-span-9nfrm.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:53 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:53 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:53 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | 2024-06-10 07:30:59 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:31:00 | examples-simple-prod-with-volumes | Deleting namespace: kuttl-test-awake-mayfly === CONT kuttl/harness/examples-business-application-injected-sidecar logger.go:42: 07:31:07 | examples-business-application-injected-sidecar | Creating namespace: kuttl-test-musical-gazelle logger.go:42: 07:31:07 | examples-business-application-injected-sidecar/0-install | starting test step 0-install logger.go:42: 07:31:07 | examples-business-application-injected-sidecar/0-install | Deployment:kuttl-test-musical-gazelle/myapp created logger.go:42: 07:31:07 | examples-business-application-injected-sidecar/0-install | test step completed 0-install logger.go:42: 07:31:07 | examples-business-application-injected-sidecar/1-install | starting test step 1-install logger.go:42: 07:31:07 | examples-business-application-injected-sidecar/1-install | Jaeger:kuttl-test-musical-gazelle/simplest created logger.go:42: 07:31:17 | examples-business-application-injected-sidecar/1-install | test step completed 1-install logger.go:42: 07:31:17 | examples-business-application-injected-sidecar/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:31:17 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:31:19 | examples-business-application-injected-sidecar/2-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:31:25 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:31:26 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:31:26 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/report-span created logger.go:42: 07:31:26 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/check-span created logger.go:42: 07:31:39 | examples-business-application-injected-sidecar/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | examples-business-application-injected-sidecar events from ns kuttl-test-musical-gazelle: logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:07 +0000 UTC Normal Pod myapp-679f79d5f8-sktzv Binding Scheduled Successfully assigned kuttl-test-musical-gazelle/myapp-679f79d5f8-sktzv to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:07 +0000 UTC Normal Pod myapp-679f79d5f8-sktzv AddedInterface Add eth0 [10.129.2.52/23] from ovn-kubernetes multus logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:07 +0000 UTC Normal Pod myapp-679f79d5f8-sktzv.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:07 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulCreate Created pod: myapp-679f79d5f8-sktzv replicaset-controller logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:07 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-679f79d5f8 to 1 deployment-controller logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:08 +0000 UTC Normal Pod myapp-856dfdf6b-9dkjf Binding Scheduled Successfully assigned kuttl-test-musical-gazelle/myapp-856dfdf6b-9dkjf to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:08 +0000 UTC Warning Pod myapp-856dfdf6b-9dkjf FailedMount MountVolume.SetUp failed for volume "simplest-trusted-ca" : configmap "simplest-trusted-ca" not found kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:08 +0000 UTC Warning Pod myapp-856dfdf6b-9dkjf FailedMount MountVolume.SetUp failed for volume "simplest-service-ca" : configmap "simplest-service-ca" not found kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:08 +0000 UTC Normal ReplicaSet.apps myapp-856dfdf6b SuccessfulCreate Created pod: myapp-856dfdf6b-9dkjf replicaset-controller logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:08 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-856dfdf6b to 1 deployment-controller logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:13 +0000 UTC Normal Pod simplest-99448444c-9mwsz Binding Scheduled Successfully assigned kuttl-test-musical-gazelle/simplest-99448444c-9mwsz to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:13 +0000 UTC Normal ReplicaSet.apps simplest-99448444c SuccessfulCreate Created pod: simplest-99448444c-9mwsz replicaset-controller logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:13 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-99448444c to 1 deployment-controller logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:14 +0000 UTC Warning Pod simplest-99448444c-9mwsz FailedMount MountVolume.SetUp failed for volume "simplest-collector-tls-config-volume" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:14 +0000 UTC Warning Pod simplest-99448444c-9mwsz FailedMount MountVolume.SetUp failed for volume "simplest-sampling-configuration-volume" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:14 +0000 UTC Warning Pod simplest-99448444c-9mwsz FailedMount MountVolume.SetUp failed for volume "kube-api-access-rjb82" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:15 +0000 UTC Normal Pod simplest-99448444c-9mwsz AddedInterface Add eth0 [10.128.2.50/23] from ovn-kubernetes multus logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:15 +0000 UTC Normal Pod simplest-99448444c-9mwsz.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:15 +0000 UTC Normal Pod simplest-99448444c-9mwsz.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:15 +0000 UTC Normal Pod simplest-99448444c-9mwsz.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:15 +0000 UTC Normal Pod simplest-99448444c-9mwsz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:15 +0000 UTC Normal Pod simplest-99448444c-9mwsz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:15 +0000 UTC Normal Pod simplest-99448444c-9mwsz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:16 +0000 UTC Normal Pod myapp-856dfdf6b-9dkjf AddedInterface Add eth0 [10.131.0.45/23] from ovn-kubernetes multus logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:16 +0000 UTC Normal Pod myapp-856dfdf6b-9dkjf.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:17 +0000 UTC Normal Pod myapp-679f79d5f8-sktzv.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 9.512s (9.512s including waiting) kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:17 +0000 UTC Normal Pod myapp-679f79d5f8-sktzv.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:17 +0000 UTC Normal Pod myapp-679f79d5f8-sktzv.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:20 +0000 UTC Normal Pod myapp-856dfdf6b-9dkjf.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 4.024s (4.024s including waiting) kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:20 +0000 UTC Normal Pod myapp-856dfdf6b-9dkjf.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:20 +0000 UTC Normal Pod myapp-856dfdf6b-9dkjf.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:20 +0000 UTC Normal Pod myapp-856dfdf6b-9dkjf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:20 +0000 UTC Normal Pod myapp-856dfdf6b-9dkjf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:20 +0000 UTC Normal Pod myapp-856dfdf6b-9dkjf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:20 +0000 UTC Normal Pod simplest-99448444c-9mwsz.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:20 +0000 UTC Normal Pod simplest-99448444c-9mwsz.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:20 +0000 UTC Normal ReplicaSet.apps simplest-99448444c SuccessfulDelete Deleted pod: simplest-99448444c-9mwsz replicaset-controller logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:20 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-99448444c to 0 from 1 deployment-controller logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:21 +0000 UTC Normal Pod myapp-679f79d5f8-sktzv.spec.containers{myapp} Killing Stopping container myapp kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:21 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulDelete Deleted pod: myapp-679f79d5f8-sktzv replicaset-controller logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:21 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled down replica set myapp-679f79d5f8 to 0 from 1 deployment-controller logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:21 +0000 UTC Normal Pod simplest-855ccc4f77-jckdt Binding Scheduled Successfully assigned kuttl-test-musical-gazelle/simplest-855ccc4f77-jckdt to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:21 +0000 UTC Normal ReplicaSet.apps simplest-855ccc4f77 SuccessfulCreate Created pod: simplest-855ccc4f77-jckdt replicaset-controller logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:21 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-855ccc4f77 to 1 deployment-controller logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:22 +0000 UTC Normal Pod simplest-855ccc4f77-jckdt AddedInterface Add eth0 [10.128.2.51/23] from ovn-kubernetes multus logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:22 +0000 UTC Normal Pod simplest-855ccc4f77-jckdt.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:22 +0000 UTC Normal Pod simplest-855ccc4f77-jckdt.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:22 +0000 UTC Normal Pod simplest-855ccc4f77-jckdt.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:22 +0000 UTC Normal Pod simplest-855ccc4f77-jckdt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:22 +0000 UTC Normal Pod simplest-855ccc4f77-jckdt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:22 +0000 UTC Normal Pod simplest-855ccc4f77-jckdt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:26 +0000 UTC Normal Pod check-span-nbml4 Binding Scheduled Successfully assigned kuttl-test-musical-gazelle/check-span-nbml4 to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:26 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-nbml4 job-controller logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:26 +0000 UTC Normal Pod report-span-7k2l7 Binding Scheduled Successfully assigned kuttl-test-musical-gazelle/report-span-7k2l7 to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:26 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-7k2l7 job-controller logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:27 +0000 UTC Normal Pod check-span-nbml4 AddedInterface Add eth0 [10.129.2.54/23] from ovn-kubernetes multus logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:27 +0000 UTC Normal Pod check-span-nbml4.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:27 +0000 UTC Normal Pod check-span-nbml4.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 334ms (334ms including waiting) kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:27 +0000 UTC Normal Pod check-span-nbml4.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:27 +0000 UTC Normal Pod check-span-nbml4.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:27 +0000 UTC Warning Pod myapp-856dfdf6b-9dkjf.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.131.0.45:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:27 +0000 UTC Normal Pod report-span-7k2l7 AddedInterface Add eth0 [10.129.2.53/23] from ovn-kubernetes multus logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:27 +0000 UTC Normal Pod report-span-7k2l7.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:27 +0000 UTC Normal Pod report-span-7k2l7.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 376ms (376ms including waiting) kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:28 +0000 UTC Normal Pod report-span-7k2l7.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:28 +0000 UTC Normal Pod report-span-7k2l7.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | 2024-06-10 07:31:38 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:31:39 | examples-business-application-injected-sidecar | Deleting namespace: kuttl-test-musical-gazelle === CONT kuttl/harness/examples-service-types logger.go:42: 07:31:51 | examples-service-types | Creating namespace: kuttl-test-united-shiner logger.go:42: 07:31:51 | examples-service-types/0-install | starting test step 0-install logger.go:42: 07:31:51 | examples-service-types/0-install | Jaeger:kuttl-test-united-shiner/service-types created logger.go:42: 07:31:58 | examples-service-types/0-install | test step completed 0-install logger.go:42: 07:31:58 | examples-service-types/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:31:58 | examples-service-types/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE service-types /dev/null] logger.go:42: 07:32:00 | examples-service-types/1-smoke-test | Warning: resource jaegers/service-types is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:32:06 | examples-service-types/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://service-types-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:32:07 | examples-service-types/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:32:07 | examples-service-types/1-smoke-test | job.batch/report-span created logger.go:42: 07:32:07 | examples-service-types/1-smoke-test | job.batch/check-span created logger.go:42: 07:32:19 | examples-service-types/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:32:19 | examples-service-types/2- | starting test step 2- logger.go:42: 07:32:20 | examples-service-types/2- | test step completed 2- logger.go:42: 07:32:20 | examples-service-types | examples-service-types events from ns kuttl-test-united-shiner: logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:31:55 +0000 UTC Normal Pod service-types-867b9d67bb-dvfjg Binding Scheduled Successfully assigned kuttl-test-united-shiner/service-types-867b9d67bb-dvfjg to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:31:55 +0000 UTC Normal Pod service-types-867b9d67bb-dvfjg AddedInterface Add eth0 [10.131.0.46/23] from ovn-kubernetes multus logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:31:55 +0000 UTC Normal Pod service-types-867b9d67bb-dvfjg.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:31:55 +0000 UTC Normal Pod service-types-867b9d67bb-dvfjg.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:31:55 +0000 UTC Normal Pod service-types-867b9d67bb-dvfjg.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:31:55 +0000 UTC Normal Pod service-types-867b9d67bb-dvfjg.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:31:55 +0000 UTC Normal ReplicaSet.apps service-types-867b9d67bb SuccessfulCreate Created pod: service-types-867b9d67bb-dvfjg replicaset-controller logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:31:55 +0000 UTC Normal Service service-types-collector EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:31:55 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-867b9d67bb to 1 deployment-controller logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:31:56 +0000 UTC Normal Pod service-types-867b9d67bb-dvfjg.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:31:56 +0000 UTC Normal Pod service-types-867b9d67bb-dvfjg.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:31:58 +0000 UTC Normal Service service-types-collector EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:31:58 +0000 UTC Normal Service service-types-query EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:00 +0000 UTC Normal Service service-types-query EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:02 +0000 UTC Normal ReplicaSet.apps service-types-85b5b5c8c8 SuccessfulCreate Created pod: service-types-85b5b5c8c8-44pl7 replicaset-controller logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:02 +0000 UTC Normal Pod service-types-867b9d67bb-dvfjg.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:02 +0000 UTC Normal Pod service-types-867b9d67bb-dvfjg.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:02 +0000 UTC Normal ReplicaSet.apps service-types-867b9d67bb SuccessfulDelete Deleted pod: service-types-867b9d67bb-dvfjg replicaset-controller logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:02 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled down replica set service-types-867b9d67bb to 0 from 1 deployment-controller logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:02 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-85b5b5c8c8 to 1 deployment-controller logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:03 +0000 UTC Normal Pod service-types-85b5b5c8c8-44pl7 Binding Scheduled Successfully assigned kuttl-test-united-shiner/service-types-85b5b5c8c8-44pl7 to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:03 +0000 UTC Normal Pod service-types-85b5b5c8c8-44pl7 AddedInterface Add eth0 [10.131.0.47/23] from ovn-kubernetes multus logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:03 +0000 UTC Normal Pod service-types-85b5b5c8c8-44pl7.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:03 +0000 UTC Normal Pod service-types-85b5b5c8c8-44pl7.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:03 +0000 UTC Normal Pod service-types-85b5b5c8c8-44pl7.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:03 +0000 UTC Normal Pod service-types-85b5b5c8c8-44pl7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:03 +0000 UTC Normal Pod service-types-85b5b5c8c8-44pl7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:03 +0000 UTC Normal Pod service-types-85b5b5c8c8-44pl7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:07 +0000 UTC Normal Pod check-span-jllxb Binding Scheduled Successfully assigned kuttl-test-united-shiner/check-span-jllxb to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:07 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-jllxb job-controller logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:07 +0000 UTC Normal Pod report-span-cp77j Binding Scheduled Successfully assigned kuttl-test-united-shiner/report-span-cp77j to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:07 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-cp77j job-controller logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:08 +0000 UTC Normal Pod check-span-jllxb AddedInterface Add eth0 [10.128.2.52/23] from ovn-kubernetes multus logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:08 +0000 UTC Normal Pod check-span-jllxb.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:08 +0000 UTC Normal Pod check-span-jllxb.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 333ms (333ms including waiting) kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:08 +0000 UTC Normal Pod check-span-jllxb.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:08 +0000 UTC Normal Pod check-span-jllxb.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:08 +0000 UTC Normal Pod report-span-cp77j AddedInterface Add eth0 [10.129.2.55/23] from ovn-kubernetes multus logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:08 +0000 UTC Normal Pod report-span-cp77j.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:08 +0000 UTC Normal Pod report-span-cp77j.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 332ms (332ms including waiting) kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:08 +0000 UTC Normal Pod report-span-cp77j.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:08 +0000 UTC Normal Pod report-span-cp77j.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:32:20 | examples-service-types | 2024-06-10 07:32:19 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:32:20 | examples-service-types | Deleting namespace: kuttl-test-united-shiner === CONT kuttl/harness/examples-openshift-with-htpasswd logger.go:42: 07:32:47 | examples-openshift-with-htpasswd | Ignoring 00-install.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:32:47 | examples-openshift-with-htpasswd | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:32:47 | examples-openshift-with-htpasswd | Creating namespace: kuttl-test-fast-marten logger.go:42: 07:32:47 | examples-openshift-with-htpasswd/0-install | starting test step 0-install logger.go:42: 07:32:47 | examples-openshift-with-htpasswd/0-install | Secret:kuttl-test-fast-marten/htpasswd created logger.go:42: 07:32:47 | examples-openshift-with-htpasswd/0-install | test step completed 0-install logger.go:42: 07:32:47 | examples-openshift-with-htpasswd/1-install | starting test step 1-install logger.go:42: 07:32:47 | examples-openshift-with-htpasswd/1-install | Jaeger:kuttl-test-fast-marten/with-htpasswd created logger.go:42: 07:32:52 | examples-openshift-with-htpasswd/1-install | test step completed 1-install logger.go:42: 07:32:52 | examples-openshift-with-htpasswd/2-check-unsecured | starting test step 2-check-unsecured logger.go:42: 07:32:52 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [./ensure-ingress-host.sh] logger.go:42: 07:32:52 | examples-openshift-with-htpasswd/2-check-unsecured | Checking the Ingress host value was populated logger.go:42: 07:32:52 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 0 logger.go:42: 07:32:52 | examples-openshift-with-htpasswd/2-check-unsecured | Hostname is with-htpasswd-kuttl-test-fast-marten.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com logger.go:42: 07:32:52 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:32:52 | examples-openshift-with-htpasswd/2-check-unsecured | Checking an expected HTTP response logger.go:42: 07:32:52 | examples-openshift-with-htpasswd/2-check-unsecured | Running in OpenShift logger.go:42: 07:32:52 | examples-openshift-with-htpasswd/2-check-unsecured | Not using any secret logger.go:42: 07:32:52 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 1/30 the https://with-htpasswd-kuttl-test-fast-marten.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 07:32:52 | examples-openshift-with-htpasswd/2-check-unsecured | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:32:52 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 2/30 the https://with-htpasswd-kuttl-test-fast-marten.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 07:32:53 | examples-openshift-with-htpasswd/2-check-unsecured | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 3/30 the https://with-htpasswd-kuttl-test-fast-marten.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/2-check-unsecured | curl response asserted properly logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/2-check-unsecured | test step completed 2-check-unsecured logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/3-check-unauthorized | starting test step 3-check-unauthorized logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking the Ingress host value was populated logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 0 logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/3-check-unauthorized | Hostname is with-htpasswd-kuttl-test-fast-marten.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [sh -c JAEGER_USERNAME=wronguser JAEGER_PASSWORD=wrongpassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking an expected HTTP response logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/3-check-unauthorized | Running in OpenShift logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/3-check-unauthorized | Using Jaeger basic authentication logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 1/30 the https://with-htpasswd-kuttl-test-fast-marten.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/3-check-unauthorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 2/30 the https://with-htpasswd-kuttl-test-fast-marten.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/3-check-unauthorized | curl response asserted properly logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/3-check-unauthorized | test step completed 3-check-unauthorized logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/4-check-authorized | starting test step 4-check-authorized logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/4-check-authorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/4-check-authorized | Checking the Ingress host value was populated logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/4-check-authorized | Try number 0 logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/4-check-authorized | Hostname is with-htpasswd-kuttl-test-fast-marten.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com logger.go:42: 07:33:03 | examples-openshift-with-htpasswd/4-check-authorized | running command: [sh -c JAEGER_USERNAME=awesomeuser JAEGER_PASSWORD=awesomepassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE with-htpasswd] logger.go:42: 07:33:04 | examples-openshift-with-htpasswd/4-check-authorized | Checking an expected HTTP response logger.go:42: 07:33:04 | examples-openshift-with-htpasswd/4-check-authorized | Running in OpenShift logger.go:42: 07:33:04 | examples-openshift-with-htpasswd/4-check-authorized | Using Jaeger basic authentication logger.go:42: 07:33:04 | examples-openshift-with-htpasswd/4-check-authorized | Try number 1/30 the https://with-htpasswd-kuttl-test-fast-marten.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 07:33:04 | examples-openshift-with-htpasswd/4-check-authorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:33:04 | examples-openshift-with-htpasswd/4-check-authorized | Try number 2/30 the https://with-htpasswd-kuttl-test-fast-marten.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 07:33:04 | examples-openshift-with-htpasswd/4-check-authorized | curl response asserted properly logger.go:42: 07:33:04 | examples-openshift-with-htpasswd/4-check-authorized | test step completed 4-check-authorized logger.go:42: 07:33:04 | examples-openshift-with-htpasswd | examples-openshift-with-htpasswd events from ns kuttl-test-fast-marten: logger.go:42: 07:33:04 | examples-openshift-with-htpasswd | 2024-06-10 07:32:50 +0000 UTC Normal Deployment.apps with-htpasswd ScalingReplicaSet Scaled up replica set with-htpasswd-748cbdf5c to 1 deployment-controller logger.go:42: 07:33:04 | examples-openshift-with-htpasswd | 2024-06-10 07:32:51 +0000 UTC Normal Pod with-htpasswd-748cbdf5c-5bpsn Binding Scheduled Successfully assigned kuttl-test-fast-marten/with-htpasswd-748cbdf5c-5bpsn to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:33:04 | examples-openshift-with-htpasswd | 2024-06-10 07:32:51 +0000 UTC Normal Pod with-htpasswd-748cbdf5c-5bpsn AddedInterface Add eth0 [10.129.2.56/23] from ovn-kubernetes multus logger.go:42: 07:33:04 | examples-openshift-with-htpasswd | 2024-06-10 07:32:51 +0000 UTC Normal Pod with-htpasswd-748cbdf5c-5bpsn.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:33:04 | examples-openshift-with-htpasswd | 2024-06-10 07:32:51 +0000 UTC Normal Pod with-htpasswd-748cbdf5c-5bpsn.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:33:04 | examples-openshift-with-htpasswd | 2024-06-10 07:32:51 +0000 UTC Normal Pod with-htpasswd-748cbdf5c-5bpsn.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:33:04 | examples-openshift-with-htpasswd | 2024-06-10 07:32:51 +0000 UTC Normal Pod with-htpasswd-748cbdf5c-5bpsn.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:33:04 | examples-openshift-with-htpasswd | 2024-06-10 07:32:51 +0000 UTC Normal Pod with-htpasswd-748cbdf5c-5bpsn.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:33:04 | examples-openshift-with-htpasswd | 2024-06-10 07:32:51 +0000 UTC Normal Pod with-htpasswd-748cbdf5c-5bpsn.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:33:04 | examples-openshift-with-htpasswd | 2024-06-10 07:32:51 +0000 UTC Normal ReplicaSet.apps with-htpasswd-748cbdf5c SuccessfulCreate Created pod: with-htpasswd-748cbdf5c-5bpsn replicaset-controller logger.go:42: 07:33:04 | examples-openshift-with-htpasswd | Deleting namespace: kuttl-test-fast-marten === CONT kuttl/harness/examples-all-in-one-with-options logger.go:42: 07:33:10 | examples-all-in-one-with-options | Creating namespace: kuttl-test-maximum-hound logger.go:42: 07:33:10 | examples-all-in-one-with-options/0-install | starting test step 0-install logger.go:42: 07:33:10 | examples-all-in-one-with-options/0-install | Jaeger:kuttl-test-maximum-hound/my-jaeger created logger.go:42: 07:33:15 | examples-all-in-one-with-options/0-install | test step completed 0-install logger.go:42: 07:33:15 | examples-all-in-one-with-options/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:33:15 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:33:17 | examples-all-in-one-with-options/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:33:23 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443/jaeger MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:33:24 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:33:24 | examples-all-in-one-with-options/1-smoke-test | job.batch/report-span created logger.go:42: 07:33:24 | examples-all-in-one-with-options/1-smoke-test | job.batch/check-span created logger.go:42: 07:33:37 | examples-all-in-one-with-options/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:33:37 | examples-all-in-one-with-options | examples-all-in-one-with-options events from ns kuttl-test-maximum-hound: logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:13 +0000 UTC Normal Pod my-jaeger-7c49fbdb5-nf2cn Binding Scheduled Successfully assigned kuttl-test-maximum-hound/my-jaeger-7c49fbdb5-nf2cn to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:13 +0000 UTC Normal ReplicaSet.apps my-jaeger-7c49fbdb5 SuccessfulCreate Created pod: my-jaeger-7c49fbdb5-nf2cn replicaset-controller logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:13 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-7c49fbdb5 to 1 deployment-controller logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:14 +0000 UTC Normal Pod my-jaeger-7c49fbdb5-nf2cn AddedInterface Add eth0 [10.131.0.48/23] from ovn-kubernetes multus logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:14 +0000 UTC Normal Pod my-jaeger-7c49fbdb5-nf2cn.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:14 +0000 UTC Normal Pod my-jaeger-7c49fbdb5-nf2cn.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:14 +0000 UTC Normal Pod my-jaeger-7c49fbdb5-nf2cn.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:14 +0000 UTC Normal Pod my-jaeger-7c49fbdb5-nf2cn.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:14 +0000 UTC Normal Pod my-jaeger-7c49fbdb5-nf2cn.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:14 +0000 UTC Normal Pod my-jaeger-7c49fbdb5-nf2cn.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:19 +0000 UTC Normal Pod my-jaeger-7c49fbdb5-nf2cn.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:19 +0000 UTC Normal Pod my-jaeger-7c49fbdb5-nf2cn.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:19 +0000 UTC Normal ReplicaSet.apps my-jaeger-7c49fbdb5 SuccessfulDelete Deleted pod: my-jaeger-7c49fbdb5-nf2cn replicaset-controller logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:19 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-7c49fbdb5 to 0 from 1 deployment-controller logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:20 +0000 UTC Normal Pod my-jaeger-87b6794fc-j8qzx Binding Scheduled Successfully assigned kuttl-test-maximum-hound/my-jaeger-87b6794fc-j8qzx to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:20 +0000 UTC Normal Pod my-jaeger-87b6794fc-j8qzx AddedInterface Add eth0 [10.128.2.53/23] from ovn-kubernetes multus logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:20 +0000 UTC Normal Pod my-jaeger-87b6794fc-j8qzx.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:20 +0000 UTC Normal Pod my-jaeger-87b6794fc-j8qzx.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:20 +0000 UTC Normal Pod my-jaeger-87b6794fc-j8qzx.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:20 +0000 UTC Normal Pod my-jaeger-87b6794fc-j8qzx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:20 +0000 UTC Normal ReplicaSet.apps my-jaeger-87b6794fc SuccessfulCreate Created pod: my-jaeger-87b6794fc-j8qzx replicaset-controller logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:20 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-87b6794fc to 1 deployment-controller logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:21 +0000 UTC Normal Pod my-jaeger-87b6794fc-j8qzx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:21 +0000 UTC Normal Pod my-jaeger-87b6794fc-j8qzx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:24 +0000 UTC Normal Pod check-span-hkgmg Binding Scheduled Successfully assigned kuttl-test-maximum-hound/check-span-hkgmg to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:24 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-hkgmg job-controller logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:24 +0000 UTC Normal Pod report-span-7dx7v Binding Scheduled Successfully assigned kuttl-test-maximum-hound/report-span-7dx7v to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:24 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-7dx7v job-controller logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:25 +0000 UTC Normal Pod check-span-hkgmg AddedInterface Add eth0 [10.131.0.49/23] from ovn-kubernetes multus logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:25 +0000 UTC Normal Pod check-span-hkgmg.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:25 +0000 UTC Normal Pod report-span-7dx7v AddedInterface Add eth0 [10.129.2.57/23] from ovn-kubernetes multus logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:25 +0000 UTC Normal Pod report-span-7dx7v.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:25 +0000 UTC Normal Pod report-span-7dx7v.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 350ms (350ms including waiting) kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:26 +0000 UTC Normal Pod check-span-hkgmg.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 980ms (980ms including waiting) kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:26 +0000 UTC Normal Pod check-span-hkgmg.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:26 +0000 UTC Normal Pod check-span-hkgmg.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:26 +0000 UTC Normal Pod report-span-7dx7v.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:26 +0000 UTC Normal Pod report-span-7dx7v.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:33:37 | examples-all-in-one-with-options | 2024-06-10 07:33:37 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:33:37 | examples-all-in-one-with-options | Deleting namespace: kuttl-test-maximum-hound === CONT kuttl/harness/examples-agent-with-priority-class logger.go:42: 07:33:49 | examples-agent-with-priority-class | Creating namespace: kuttl-test-central-hermit logger.go:42: 07:33:49 | examples-agent-with-priority-class/0-install | starting test step 0-install logger.go:42: 07:33:49 | examples-agent-with-priority-class/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 07:33:49 | examples-agent-with-priority-class/0-install | ServiceAccount:kuttl-test-central-hermit/jaeger-agent-daemonset created logger.go:42: 07:33:49 | examples-agent-with-priority-class/0-install | test step completed 0-install logger.go:42: 07:33:49 | examples-agent-with-priority-class/1-add-policy | starting test step 1-add-policy logger.go:42: 07:33:49 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 07:33:49 | examples-agent-with-priority-class/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 07:33:49 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 07:33:54 | examples-agent-with-priority-class/1-add-policy | test step completed 1-add-policy logger.go:42: 07:33:54 | examples-agent-with-priority-class/2-install | starting test step 2-install logger.go:42: 07:33:54 | examples-agent-with-priority-class/2-install | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 07:34:03 | examples-agent-with-priority-class/2-install | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:34:03 | examples-agent-with-priority-class/2-install | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:34:04 | examples-agent-with-priority-class/2-install | job.batch/report-span created logger.go:42: 07:34:04 | examples-agent-with-priority-class/2-install | job.batch/check-span created logger.go:42: 07:34:04 | examples-agent-with-priority-class/2-install | PriorityClass:/high-priority created logger.go:42: 07:34:04 | examples-agent-with-priority-class/2-install | Jaeger:kuttl-test-central-hermit/agent-as-daemonset updated logger.go:42: 07:34:15 | examples-agent-with-priority-class/2-install | test step completed 2-install logger.go:42: 07:34:15 | examples-agent-with-priority-class | examples-agent-with-priority-class events from ns kuttl-test-central-hermit: logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:00 +0000 UTC Normal Pod agent-as-daemonset-69ccd6d5d6-gk95z Binding Scheduled Successfully assigned kuttl-test-central-hermit/agent-as-daemonset-69ccd6d5d6-gk95z to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:00 +0000 UTC Warning Pod agent-as-daemonset-69ccd6d5d6-gk95z FailedMount MountVolume.SetUp failed for volume "agent-as-daemonset-collector-tls-config-volume" : secret "agent-as-daemonset-collector-headless-tls" not found kubelet logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:00 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-69ccd6d5d6 SuccessfulCreate Created pod: agent-as-daemonset-69ccd6d5d6-gk95z replicaset-controller logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:00 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-69ccd6d5d6 to 1 deployment-controller logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:01 +0000 UTC Normal Pod agent-as-daemonset-69ccd6d5d6-gk95z AddedInterface Add eth0 [10.131.0.50/23] from ovn-kubernetes multus logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:01 +0000 UTC Normal Pod agent-as-daemonset-69ccd6d5d6-gk95z.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:01 +0000 UTC Normal Pod agent-as-daemonset-69ccd6d5d6-gk95z.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:01 +0000 UTC Normal Pod agent-as-daemonset-69ccd6d5d6-gk95z.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:01 +0000 UTC Normal Pod agent-as-daemonset-69ccd6d5d6-gk95z.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:01 +0000 UTC Normal Pod agent-as-daemonset-69ccd6d5d6-gk95z.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:01 +0000 UTC Normal Pod agent-as-daemonset-69ccd6d5d6-gk95z.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:04 +0000 UTC Normal Pod check-span-lqgtt Binding Scheduled Successfully assigned kuttl-test-central-hermit/check-span-lqgtt to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:04 +0000 UTC Normal Pod check-span-lqgtt AddedInterface Add eth0 [10.128.2.54/23] from ovn-kubernetes multus logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:04 +0000 UTC Normal Pod check-span-lqgtt.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:04 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-lqgtt job-controller logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:04 +0000 UTC Normal Pod report-span-fw56d Binding Scheduled Successfully assigned kuttl-test-central-hermit/report-span-fw56d to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:04 +0000 UTC Normal Pod report-span-fw56d AddedInterface Add eth0 [10.129.2.58/23] from ovn-kubernetes multus logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:04 +0000 UTC Normal Pod report-span-fw56d.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:04 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-fw56d job-controller logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:05 +0000 UTC Warning DaemonSet.apps agent-as-daemonset-agent-daemonset FailedCreate Error creating: pods "agent-as-daemonset-agent-daemonset-" is forbidden: unable to validate against any security context constraint: [provider "anyuid": Forbidden: not usable by user or serviceaccount, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5775: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5778: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6831: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6832: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 14271: Host ports are not allowed to be used, provider "restricted": Forbidden: not usable by user or serviceaccount, provider "nonroot-v2": Forbidden: not usable by user or serviceaccount, provider "nonroot": Forbidden: not usable by user or serviceaccount, provider "hostmount-anyuid": Forbidden: not usable by user or serviceaccount, provider "elasticsearch-scc": Forbidden: not usable by user or serviceaccount, provider "machine-api-termination-handler": Forbidden: not usable by user or serviceaccount, provider "daemonset-with-hostport": Forbidden: not usable by user or serviceaccount, provider "hostnetwork-v2": Forbidden: not usable by user or serviceaccount, provider "hostnetwork": Forbidden: not usable by user or serviceaccount, provider "hostaccess": Forbidden: not usable by user or serviceaccount, provider "node-exporter": Forbidden: not usable by user or serviceaccount, provider "privileged": Forbidden: not usable by user or serviceaccount] daemonset-controller logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:05 +0000 UTC Normal Pod check-span-lqgtt.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 346ms (346ms including waiting) kubelet logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:05 +0000 UTC Normal Pod check-span-lqgtt.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:05 +0000 UTC Normal Pod check-span-lqgtt.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:05 +0000 UTC Normal Pod report-span-fw56d.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 461ms (461ms including waiting) kubelet logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:05 +0000 UTC Normal Pod report-span-fw56d.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:05 +0000 UTC Normal Pod report-span-fw56d.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:34:15 | examples-agent-with-priority-class | 2024-06-10 07:34:15 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:34:15 | examples-agent-with-priority-class | Deleting namespace: kuttl-test-central-hermit === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (921.88s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.05s) --- PASS: kuttl/harness/examples-simple-prod (63.81s) --- PASS: kuttl/harness/examples-with-sampling (345.73s) --- PASS: kuttl/harness/examples-with-cassandra (150.83s) --- PASS: kuttl/harness/examples-with-badger (39.60s) --- PASS: kuttl/harness/examples-simplest (41.44s) --- PASS: kuttl/harness/examples-simple-prod-with-volumes (66.38s) --- PASS: kuttl/harness/examples-business-application-injected-sidecar (44.18s) --- PASS: kuttl/harness/examples-service-types (55.87s) --- PASS: kuttl/harness/examples-openshift-with-htpasswd (22.97s) --- PASS: kuttl/harness/examples-all-in-one-with-options (39.18s) --- PASS: kuttl/harness/examples-agent-with-priority-class (45.63s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name examples --report --output /logs/artifacts/examples.xml ./artifacts/kuttl-report.xml time="2024-06-10T07:34:35Z" level=debug msg="Setting a new name for the test suites" time="2024-06-10T07:34:35Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-06-10T07:34:35Z" level=debug msg="normalizing test case names" time="2024-06-10T07:34:35Z" level=debug msg="examples/artifacts -> examples_artifacts" time="2024-06-10T07:34:35Z" level=debug msg="examples/examples-simple-prod -> examples_examples_simple_prod" time="2024-06-10T07:34:35Z" level=debug msg="examples/examples-with-sampling -> examples_examples_with_sampling" time="2024-06-10T07:34:35Z" level=debug msg="examples/examples-with-cassandra -> examples_examples_with_cassandra" time="2024-06-10T07:34:35Z" level=debug msg="examples/examples-with-badger -> examples_examples_with_badger" time="2024-06-10T07:34:35Z" level=debug msg="examples/examples-simplest -> examples_examples_simplest" time="2024-06-10T07:34:35Z" level=debug msg="examples/examples-simple-prod-with-volumes -> examples_examples_simple_prod_with_volumes" time="2024-06-10T07:34:35Z" level=debug msg="examples/examples-business-application-injected-sidecar -> examples_examples_business_application_injected_sidecar" time="2024-06-10T07:34:35Z" level=debug msg="examples/examples-service-types -> examples_examples_service_types" time="2024-06-10T07:34:35Z" level=debug msg="examples/examples-openshift-with-htpasswd -> examples_examples_openshift_with_htpasswd" time="2024-06-10T07:34:35Z" level=debug msg="examples/examples-all-in-one-with-options -> examples_examples_all_in_one_with_options" time="2024-06-10T07:34:35Z" level=debug msg="examples/examples-agent-with-priority-class -> examples_examples_agent_with_priority_class" +---------------------------------------------------------+--------+ | NAME | RESULT | +---------------------------------------------------------+--------+ | examples_artifacts | passed | | examples_examples_simple_prod | passed | | examples_examples_with_sampling | passed | | examples_examples_with_cassandra | passed | | examples_examples_with_badger | passed | | examples_examples_simplest | passed | | examples_examples_simple_prod_with_volumes | passed | | examples_examples_business_application_injected_sidecar | passed | | examples_examples_service_types | passed | | examples_examples_openshift_with_htpasswd | passed | | examples_examples_all_in_one_with_options | passed | | examples_examples_agent_with_priority_class | passed | +---------------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh generate false true + '[' 3 -ne 3 ']' + test_suite_name=generate + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/generate.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-generate make[2]: Entering directory '/tmp/jaeger-tests' test -s /tmp/jaeger-tests/bin/operator-sdk || curl -sLo /tmp/jaeger-tests/bin/operator-sdk https://github.com/operator-framework/operator-sdk/releases/download/v1.32.0/operator-sdk_`go env GOOS`_`go env GOARCH` ./hack/install/install-golangci-lint.sh Installing golangci-lint golangci-lint 1.55.2 is installed already ./hack/install/install-goimports.sh Installing goimports Try 0... go install golang.org/x/tools/cmd/goimports@v0.1.12 >>>> Formatting code... ./.ci/format.sh >>>> Building... ./hack/install/install-dependencies.sh Installing go dependencies Try 0... go mod download GOOS= GOARCH= CGO_ENABLED=0 GO111MODULE=on go build -ldflags "-X "github.com/jaegertracing/jaeger-operator/pkg/version".version="1.57.0" -X "github.com/jaegertracing/jaeger-operator/pkg/version".buildDate=2024-06-10T07:34:37Z -X "github.com/jaegertracing/jaeger-operator/pkg/version".defaultJaeger="1.57.0"" -o "bin/jaeger-operator" main.go JAEGER_VERSION="1.57.0" ./tests/e2e/generate/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-07-024651 True False 48m Cluster version is 4.16.0-0.nightly-2024-06-07-024651' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-07-024651 True False 48m Cluster version is 4.16.0-0.nightly-2024-06-07-024651' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/generate/render.sh ++ export SUITE_DIR=./tests/e2e/generate ++ SUITE_DIR=./tests/e2e/generate ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/generate ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test generate 'This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 2 -ne 2 ']' + test_name=generate + message='This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/generate/_build + '[' _build '!=' _build ']' + rm -rf generate + warning 'generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed\e[0m' WAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running generate E2E tests' Running generate E2E tests + cd tests/e2e/generate/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3357233117 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:34:53 | artifacts | Creating namespace: kuttl-test-peaceful-meerkat logger.go:42: 07:34:53 | artifacts | artifacts events from ns kuttl-test-peaceful-meerkat: logger.go:42: 07:34:53 | artifacts | Deleting namespace: kuttl-test-peaceful-meerkat === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.15s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.94s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name generate --report --output /logs/artifacts/generate.xml ./artifacts/kuttl-report.xml time="2024-06-10T07:34:59Z" level=debug msg="Setting a new name for the test suites" time="2024-06-10T07:34:59Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-06-10T07:34:59Z" level=debug msg="normalizing test case names" time="2024-06-10T07:34:59Z" level=debug msg="generate/artifacts -> generate_artifacts" +--------------------+--------+ | NAME | RESULT | +--------------------+--------+ | generate_artifacts | passed | +--------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh upgrade false true + '[' 3 -ne 3 ']' + test_suite_name=upgrade + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/upgrade.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-upgrade make[2]: Entering directory '/tmp/jaeger-tests' make docker JAEGER_VERSION=1.57.1 IMG="quay.io//jaeger-operator:next" make[3]: Entering directory '/tmp/jaeger-tests' [ ! -z "true" ] || docker build --build-arg=GOPROXY= --build-arg=VERSION="1.57.0" --build-arg=JAEGER_VERSION=1.57.1 --build-arg=TARGETARCH= --build-arg VERSION_DATE=2024-06-10T07:34:59Z --build-arg VERSION_PKG="github.com/jaegertracing/jaeger-operator/pkg/version" -t "quay.io//jaeger-operator:next" . make[3]: Leaving directory '/tmp/jaeger-tests' touch build-e2e-upgrade-image SKIP_ES_EXTERNAL=true IMG=quay.io//jaeger-operator:"1.57.0" JAEGER_OPERATOR_VERSION="1.57.0" JAEGER_VERSION="1.57.0" ./tests/e2e/upgrade/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-07-024651 True False 48m Cluster version is 4.16.0-0.nightly-2024-06-07-024651' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-07-024651 True False 48m Cluster version is 4.16.0-0.nightly-2024-06-07-024651' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/upgrade/render.sh ++ export SUITE_DIR=./tests/e2e/upgrade ++ SUITE_DIR=./tests/e2e/upgrade ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/upgrade ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + export JAEGER_NAME + '[' true = true ']' + skip_test upgrade 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade + warning 'upgrade: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade: Test not supported in OpenShift\e[0m' WAR: upgrade: Test not supported in OpenShift + '[' true = true ']' + skip_test upgrade-from-latest-release 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade-from-latest-release + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade-from-latest-release + warning 'upgrade-from-latest-release: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade-from-latest-release: Test not supported in OpenShift\e[0m' WAR: upgrade-from-latest-release: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running upgrade E2E tests' Running upgrade E2E tests + cd tests/e2e/upgrade/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3357233117 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:35:00 | artifacts | Creating namespace: kuttl-test-adjusted-cicada logger.go:42: 07:35:00 | artifacts | artifacts events from ns kuttl-test-adjusted-cicada: logger.go:42: 07:35:00 | artifacts | Deleting namespace: kuttl-test-adjusted-cicada === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.14s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.92s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name upgrade --report --output /logs/artifacts/upgrade.xml ./artifacts/kuttl-report.xml time="2024-06-10T07:35:06Z" level=debug msg="Setting a new name for the test suites" time="2024-06-10T07:35:06Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-06-10T07:35:06Z" level=debug msg="normalizing test case names" time="2024-06-10T07:35:06Z" level=debug msg="upgrade/artifacts -> upgrade_artifacts" +-------------------+--------+ | NAME | RESULT | +-------------------+--------+ | upgrade_artifacts | passed | +-------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh sidecar false true + '[' 3 -ne 3 ']' + test_suite_name=sidecar + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/sidecar.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-sidecar make[2]: Entering directory '/tmp/jaeger-tests' ./tests/e2e/sidecar/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-07-024651 True False 48m Cluster version is 4.16.0-0.nightly-2024-06-07-024651' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-07-024651 True False 48m Cluster version is 4.16.0-0.nightly-2024-06-07-024651' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/sidecar/render.sh ++ export SUITE_DIR=./tests/e2e/sidecar ++ SUITE_DIR=./tests/e2e/sidecar ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/sidecar ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + jaeger_service_name=order + start_test sidecar-deployment + '[' 1 -ne 1 ']' + test_name=sidecar-deployment + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-deployment' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-deployment\e[0m' Rendering files for test sidecar-deployment + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build + '[' _build '!=' _build ']' + mkdir -p sidecar-deployment + cd sidecar-deployment + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-namespace + '[' 1 -ne 1 ']' + test_name=sidecar-namespace + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-namespace' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-namespace\e[0m' Rendering files for test sidecar-namespace + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-deployment + '[' sidecar-deployment '!=' _build ']' + cd .. + mkdir -p sidecar-namespace + cd sidecar-namespace + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-skip-webhook + '[' 1 -ne 1 ']' + test_name=sidecar-skip-webhook + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-skip-webhook' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-skip-webhook\e[0m' Rendering files for test sidecar-skip-webhook + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-namespace + '[' sidecar-namespace '!=' _build ']' + cd .. + mkdir -p sidecar-skip-webhook + cd sidecar-skip-webhook + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running sidecar E2E tests' Running sidecar E2E tests + cd tests/e2e/sidecar/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3357233117 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/sidecar-deployment === PAUSE kuttl/harness/sidecar-deployment === RUN kuttl/harness/sidecar-namespace === PAUSE kuttl/harness/sidecar-namespace === RUN kuttl/harness/sidecar-skip-webhook === PAUSE kuttl/harness/sidecar-skip-webhook === CONT kuttl/harness/artifacts logger.go:42: 07:35:14 | artifacts | Creating namespace: kuttl-test-sweet-crayfish logger.go:42: 07:35:15 | artifacts | artifacts events from ns kuttl-test-sweet-crayfish: logger.go:42: 07:35:15 | artifacts | Deleting namespace: kuttl-test-sweet-crayfish === CONT kuttl/harness/sidecar-namespace logger.go:42: 07:35:20 | sidecar-namespace | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:35:20 | sidecar-namespace | Creating namespace: kuttl-test-faithful-pup logger.go:42: 07:35:20 | sidecar-namespace/0-install | starting test step 0-install logger.go:42: 07:35:21 | sidecar-namespace/0-install | Jaeger:kuttl-test-faithful-pup/agent-as-sidecar created logger.go:42: 07:35:27 | sidecar-namespace/0-install | test step completed 0-install logger.go:42: 07:35:27 | sidecar-namespace/1-install | starting test step 1-install logger.go:42: 07:35:27 | sidecar-namespace/1-install | Deployment:kuttl-test-faithful-pup/vertx-create-span-sidecar created logger.go:42: 07:35:29 | sidecar-namespace/1-install | test step completed 1-install logger.go:42: 07:35:29 | sidecar-namespace/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:35:29 | sidecar-namespace/2-enable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="true"] logger.go:42: 07:35:29 | sidecar-namespace/2-enable-injection | namespace/kuttl-test-faithful-pup annotated logger.go:42: 07:35:34 | sidecar-namespace/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:35:34 | sidecar-namespace/3-find-service | starting test step 3-find-service logger.go:42: 07:35:34 | sidecar-namespace/3-find-service | Job:kuttl-test-faithful-pup/00-find-service created logger.go:42: 07:35:47 | sidecar-namespace/3-find-service | test step completed 3-find-service logger.go:42: 07:35:47 | sidecar-namespace/4-other-instance | starting test step 4-other-instance logger.go:42: 07:35:47 | sidecar-namespace/4-other-instance | Jaeger:kuttl-test-faithful-pup/agent-as-sidecar2 created logger.go:42: 07:35:56 | sidecar-namespace/4-other-instance | test step completed 4-other-instance logger.go:42: 07:35:56 | sidecar-namespace/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:35:56 | sidecar-namespace/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:35:56 | sidecar-namespace/6-find-service | starting test step 6-find-service logger.go:42: 07:35:56 | sidecar-namespace/6-find-service | Job:kuttl-test-faithful-pup/01-find-service created logger.go:42: 07:36:16 | sidecar-namespace/6-find-service | test step completed 6-find-service logger.go:42: 07:36:16 | sidecar-namespace/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:36:16 | sidecar-namespace/7-disable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="false"] logger.go:42: 07:36:16 | sidecar-namespace/7-disable-injection | namespace/kuttl-test-faithful-pup annotated logger.go:42: 07:36:18 | sidecar-namespace/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:36:18 | sidecar-namespace | sidecar-namespace events from ns kuttl-test-faithful-pup: logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:24 +0000 UTC Normal Pod agent-as-sidecar-f7c697689-kh5zb Binding Scheduled Successfully assigned kuttl-test-faithful-pup/agent-as-sidecar-f7c697689-kh5zb to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:24 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-f7c697689 SuccessfulCreate Created pod: agent-as-sidecar-f7c697689-kh5zb replicaset-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:24 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-f7c697689 to 1 deployment-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:25 +0000 UTC Normal Pod agent-as-sidecar-f7c697689-kh5zb AddedInterface Add eth0 [10.131.0.51/23] from ovn-kubernetes multus logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:25 +0000 UTC Normal Pod agent-as-sidecar-f7c697689-kh5zb.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:25 +0000 UTC Normal Pod agent-as-sidecar-f7c697689-kh5zb.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:25 +0000 UTC Normal Pod agent-as-sidecar-f7c697689-kh5zb.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:27 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-j82xq Binding Scheduled Successfully assigned kuttl-test-faithful-pup/vertx-create-span-sidecar-84d458b68c-j82xq to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:27 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-j82xq AddedInterface Add eth0 [10.129.2.59/23] from ovn-kubernetes multus logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:27 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-j82xq.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:27 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-j82xq replicaset-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:27 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:28 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-j82xq.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:28 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-j82xq.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:29 +0000 UTC Normal Pod vertx-create-span-sidecar-656758476d-c4plz Binding Scheduled Successfully assigned kuttl-test-faithful-pup/vertx-create-span-sidecar-656758476d-c4plz to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:29 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-656758476d SuccessfulCreate Created pod: vertx-create-span-sidecar-656758476d-c4plz replicaset-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:29 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-656758476d to 1 deployment-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:30 +0000 UTC Normal Pod vertx-create-span-sidecar-656758476d-c4plz AddedInterface Add eth0 [10.128.2.55/23] from ovn-kubernetes multus logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:30 +0000 UTC Normal Pod vertx-create-span-sidecar-656758476d-c4plz.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:34 +0000 UTC Normal Pod vertx-create-span-sidecar-656758476d-c4plz.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.736s (3.736s including waiting) kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:34 +0000 UTC Normal Pod vertx-create-span-sidecar-656758476d-c4plz.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:34 +0000 UTC Normal Pod vertx-create-span-sidecar-656758476d-c4plz.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:34 +0000 UTC Normal Pod vertx-create-span-sidecar-656758476d-c4plz.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:34 +0000 UTC Normal Pod vertx-create-span-sidecar-656758476d-c4plz.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:34 +0000 UTC Normal Pod vertx-create-span-sidecar-656758476d-c4plz.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:35 +0000 UTC Normal Pod 00-find-service-xp5fv Binding Scheduled Successfully assigned kuttl-test-faithful-pup/00-find-service-xp5fv to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:35 +0000 UTC Normal Pod 00-find-service-xp5fv AddedInterface Add eth0 [10.131.0.52/23] from ovn-kubernetes multus logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:35 +0000 UTC Normal Pod 00-find-service-xp5fv.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:35 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-xp5fv job-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:36 +0000 UTC Normal Pod 00-find-service-xp5fv.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 371ms (372ms including waiting) kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:36 +0000 UTC Normal Pod 00-find-service-xp5fv.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:36 +0000 UTC Normal Pod 00-find-service-xp5fv.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:36 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-j82xq.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.59:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:36 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-j82xq.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.59:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:38 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-j82xq.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:39 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-j82xq.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.59:8080/": read tcp 10.129.2.2:35570->10.129.2.59:8080: read: connection reset by peer kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:39 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-j82xq.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.59:8080/": dial tcp 10.129.2.59:8080: connect: connection refused kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:43 +0000 UTC Warning Pod vertx-create-span-sidecar-656758476d-c4plz.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.55:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:43 +0000 UTC Warning Pod vertx-create-span-sidecar-656758476d-c4plz.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.55:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:45 +0000 UTC Normal Pod vertx-create-span-sidecar-656758476d-c4plz.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:45 +0000 UTC Warning Pod vertx-create-span-sidecar-656758476d-c4plz.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.55:8080/": read tcp 10.128.2.2:35646->10.128.2.55:8080: read: connection reset by peer kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:45 +0000 UTC Warning Pod vertx-create-span-sidecar-656758476d-c4plz.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.55:8080/": dial tcp 10.128.2.55:8080: connect: connection refused kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:45 +0000 UTC Normal Pod vertx-create-span-sidecar-656758476d-c4plz.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:46 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:50 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-j82xq.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.59:8080/": read tcp 10.129.2.2:39706->10.129.2.59:8080: read: connection reset by peer kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:53 +0000 UTC Normal Pod agent-as-sidecar2-5d95547dfc-6xmzf Binding Scheduled Successfully assigned kuttl-test-faithful-pup/agent-as-sidecar2-5d95547dfc-6xmzf to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:53 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-5d95547dfc SuccessfulCreate Created pod: agent-as-sidecar2-5d95547dfc-6xmzf replicaset-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:53 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-5d95547dfc to 1 deployment-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:54 +0000 UTC Normal Pod agent-as-sidecar2-5d95547dfc-6xmzf AddedInterface Add eth0 [10.131.0.53/23] from ovn-kubernetes multus logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:54 +0000 UTC Normal Pod agent-as-sidecar2-5d95547dfc-6xmzf.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:54 +0000 UTC Normal Pod agent-as-sidecar2-5d95547dfc-6xmzf.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:54 +0000 UTC Normal Pod agent-as-sidecar2-5d95547dfc-6xmzf.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:56 +0000 UTC Normal Pod 01-find-service-7lz7f Binding Scheduled Successfully assigned kuttl-test-faithful-pup/01-find-service-7lz7f to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:56 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-7lz7f job-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:56 +0000 UTC Normal Pod agent-as-sidecar-f7c697689-kh5zb.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:57 +0000 UTC Normal Pod 01-find-service-7lz7f AddedInterface Add eth0 [10.129.2.60/23] from ovn-kubernetes multus logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:57 +0000 UTC Normal Pod 01-find-service-7lz7f.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:57 +0000 UTC Normal Pod 01-find-service-7lz7f.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 353ms (353ms including waiting) kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:57 +0000 UTC Normal Pod 01-find-service-7lz7f.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:35:57 +0000 UTC Normal Pod 01-find-service-7lz7f.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:01 +0000 UTC Normal Pod vertx-create-span-sidecar-5765d79b86-j7w6b Binding Scheduled Successfully assigned kuttl-test-faithful-pup/vertx-create-span-sidecar-5765d79b86-j7w6b to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:01 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-5765d79b86 SuccessfulCreate Created pod: vertx-create-span-sidecar-5765d79b86-j7w6b replicaset-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:01 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-j82xq replicaset-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:01 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:01 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-5765d79b86 to 1 from 0 deployment-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:02 +0000 UTC Normal Pod vertx-create-span-sidecar-5765d79b86-j7w6b AddedInterface Add eth0 [10.131.0.54/23] from ovn-kubernetes multus logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:02 +0000 UTC Normal Pod vertx-create-span-sidecar-5765d79b86-j7w6b.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:02 +0000 UTC Normal Pod vertx-create-span-sidecar-5765d79b86-j7w6b.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:02 +0000 UTC Normal Pod vertx-create-span-sidecar-5765d79b86-j7w6b.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:02 +0000 UTC Normal Pod vertx-create-span-sidecar-5765d79b86-j7w6b.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:02 +0000 UTC Normal Pod vertx-create-span-sidecar-5765d79b86-j7w6b.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:02 +0000 UTC Normal Pod vertx-create-span-sidecar-5765d79b86-j7w6b.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:10 +0000 UTC Warning Pod vertx-create-span-sidecar-5765d79b86-j7w6b.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.54:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:10 +0000 UTC Warning Pod vertx-create-span-sidecar-5765d79b86-j7w6b.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.54:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:13 +0000 UTC Normal Pod vertx-create-span-sidecar-5765d79b86-j7w6b.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:13 +0000 UTC Warning Pod vertx-create-span-sidecar-5765d79b86-j7w6b.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.54:8080/": read tcp 10.131.0.2:55938->10.131.0.54:8080: read: connection reset by peer kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:13 +0000 UTC Warning Pod vertx-create-span-sidecar-5765d79b86-j7w6b.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.54:8080/": dial tcp 10.131.0.54:8080: connect: connection refused kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:16 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:16 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-656758476d SuccessfulDelete Deleted pod: vertx-create-span-sidecar-656758476d-c4plz replicaset-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:16 +0000 UTC Normal Pod vertx-create-span-sidecar-d9db5f695-vggpm Binding Scheduled Successfully assigned kuttl-test-faithful-pup/vertx-create-span-sidecar-d9db5f695-vggpm to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:16 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-d9db5f695 SuccessfulCreate Created pod: vertx-create-span-sidecar-d9db5f695-vggpm replicaset-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:16 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-656758476d to 0 from 1 deployment-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:16 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-d9db5f695 to 1 from 0 deployment-controller logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:17 +0000 UTC Normal Pod vertx-create-span-sidecar-d9db5f695-vggpm AddedInterface Add eth0 [10.129.2.61/23] from ovn-kubernetes multus logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:17 +0000 UTC Normal Pod vertx-create-span-sidecar-d9db5f695-vggpm.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:17 +0000 UTC Normal Pod vertx-create-span-sidecar-d9db5f695-vggpm.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:36:18 | sidecar-namespace | 2024-06-10 07:36:17 +0000 UTC Normal Pod vertx-create-span-sidecar-d9db5f695-vggpm.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:36:18 | sidecar-namespace | Deleting namespace: kuttl-test-faithful-pup === CONT kuttl/harness/sidecar-skip-webhook logger.go:42: 07:36:25 | sidecar-skip-webhook | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:36:25 | sidecar-skip-webhook | Creating namespace: kuttl-test-discrete-seahorse logger.go:42: 07:36:25 | sidecar-skip-webhook/0-install | starting test step 0-install logger.go:42: 07:36:25 | sidecar-skip-webhook/0-install | Jaeger:kuttl-test-discrete-seahorse/agent-as-sidecar created logger.go:42: 07:36:30 | sidecar-skip-webhook/0-install | test step completed 0-install logger.go:42: 07:36:30 | sidecar-skip-webhook/1-install | starting test step 1-install logger.go:42: 07:36:30 | sidecar-skip-webhook/1-install | Deployment:kuttl-test-discrete-seahorse/vertx-create-span-sidecar created logger.go:42: 07:36:32 | sidecar-skip-webhook/1-install | test step completed 1-install logger.go:42: 07:36:32 | sidecar-skip-webhook/2-add-anotation-and-label | starting test step 2-add-anotation-and-label logger.go:42: 07:36:32 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name=jaeger-operator --namespace kuttl-test-discrete-seahorse] logger.go:42: 07:36:32 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar labeled logger.go:42: 07:36:32 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-discrete-seahorse] logger.go:42: 07:36:32 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:36:33 | sidecar-skip-webhook/2-add-anotation-and-label | test step completed 2-add-anotation-and-label logger.go:42: 07:36:33 | sidecar-skip-webhook/3-remove-label | starting test step 3-remove-label logger.go:42: 07:36:33 | sidecar-skip-webhook/3-remove-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name- --namespace kuttl-test-discrete-seahorse] logger.go:42: 07:36:33 | sidecar-skip-webhook/3-remove-label | deployment.apps/vertx-create-span-sidecar unlabeled logger.go:42: 07:36:35 | sidecar-skip-webhook/3-remove-label | test step completed 3-remove-label logger.go:42: 07:36:35 | sidecar-skip-webhook | sidecar-skip-webhook events from ns kuttl-test-discrete-seahorse: logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:28 +0000 UTC Normal Pod agent-as-sidecar-6c489d96d8-pllxr Binding Scheduled Successfully assigned kuttl-test-discrete-seahorse/agent-as-sidecar-6c489d96d8-pllxr to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:28 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-6c489d96d8 SuccessfulCreate Created pod: agent-as-sidecar-6c489d96d8-pllxr replicaset-controller logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:28 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-6c489d96d8 to 1 deployment-controller logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:29 +0000 UTC Normal Pod agent-as-sidecar-6c489d96d8-pllxr AddedInterface Add eth0 [10.129.2.62/23] from ovn-kubernetes multus logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:29 +0000 UTC Normal Pod agent-as-sidecar-6c489d96d8-pllxr.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:29 +0000 UTC Normal Pod agent-as-sidecar-6c489d96d8-pllxr.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:29 +0000 UTC Normal Pod agent-as-sidecar-6c489d96d8-pllxr.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:30 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-flw9z Binding Scheduled Successfully assigned kuttl-test-discrete-seahorse/vertx-create-span-sidecar-84d458b68c-flw9z to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:30 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-flw9z replicaset-controller logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:30 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:31 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-flw9z AddedInterface Add eth0 [10.131.0.55/23] from ovn-kubernetes multus logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:31 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-flw9z.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:31 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-flw9z.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:31 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-flw9z.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:32 +0000 UTC Normal Pod vertx-create-span-sidecar-76cddd8c5c-r9zdl Binding Scheduled Successfully assigned kuttl-test-discrete-seahorse/vertx-create-span-sidecar-76cddd8c5c-r9zdl to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:32 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-76cddd8c5c SuccessfulCreate Created pod: vertx-create-span-sidecar-76cddd8c5c-r9zdl replicaset-controller logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:32 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-76cddd8c5c to 1 deployment-controller logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:33 +0000 UTC Normal Pod vertx-create-span-sidecar-76cddd8c5c-r9zdl AddedInterface Add eth0 [10.128.2.56/23] from ovn-kubernetes multus logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:33 +0000 UTC Normal Pod vertx-create-span-sidecar-76cddd8c5c-r9zdl.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:33 +0000 UTC Normal Pod vertx-create-span-sidecar-76cddd8c5c-r9zdl.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:33 +0000 UTC Normal Pod vertx-create-span-sidecar-76cddd8c5c-r9zdl.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:33 +0000 UTC Normal Pod vertx-create-span-sidecar-76cddd8c5c-r9zdl.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:33 +0000 UTC Normal Pod vertx-create-span-sidecar-76cddd8c5c-r9zdl.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:36:35 | sidecar-skip-webhook | 2024-06-10 07:36:33 +0000 UTC Normal Pod vertx-create-span-sidecar-76cddd8c5c-r9zdl.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:36:35 | sidecar-skip-webhook | Deleting namespace: kuttl-test-discrete-seahorse === CONT kuttl/harness/sidecar-deployment logger.go:42: 07:36:42 | sidecar-deployment | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:36:42 | sidecar-deployment | Creating namespace: kuttl-test-becoming-kite logger.go:42: 07:36:42 | sidecar-deployment/0-install | starting test step 0-install logger.go:42: 07:36:42 | sidecar-deployment/0-install | Jaeger:kuttl-test-becoming-kite/agent-as-sidecar created logger.go:42: 07:36:48 | sidecar-deployment/0-install | test step completed 0-install logger.go:42: 07:36:48 | sidecar-deployment/1-install | starting test step 1-install logger.go:42: 07:36:48 | sidecar-deployment/1-install | Deployment:kuttl-test-becoming-kite/vertx-create-span-sidecar created logger.go:42: 07:36:51 | sidecar-deployment/1-install | test step completed 1-install logger.go:42: 07:36:51 | sidecar-deployment/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:36:51 | sidecar-deployment/2-enable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-becoming-kite] logger.go:42: 07:36:51 | sidecar-deployment/2-enable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:36:53 | sidecar-deployment/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:36:53 | sidecar-deployment/3-find-service | starting test step 3-find-service logger.go:42: 07:36:53 | sidecar-deployment/3-find-service | Job:kuttl-test-becoming-kite/00-find-service created logger.go:42: 07:37:05 | sidecar-deployment/3-find-service | test step completed 3-find-service logger.go:42: 07:37:05 | sidecar-deployment/4-other-instance | starting test step 4-other-instance logger.go:42: 07:37:05 | sidecar-deployment/4-other-instance | Jaeger:kuttl-test-becoming-kite/agent-as-sidecar2 created logger.go:42: 07:37:12 | sidecar-deployment/4-other-instance | test step completed 4-other-instance logger.go:42: 07:37:12 | sidecar-deployment/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:37:12 | sidecar-deployment/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:37:12 | sidecar-deployment/6-find-service | starting test step 6-find-service logger.go:42: 07:37:12 | sidecar-deployment/6-find-service | Job:kuttl-test-becoming-kite/01-find-service created logger.go:42: 07:37:32 | sidecar-deployment/6-find-service | test step completed 6-find-service logger.go:42: 07:37:32 | sidecar-deployment/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:37:32 | sidecar-deployment/7-disable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=false --namespace kuttl-test-becoming-kite] logger.go:42: 07:37:32 | sidecar-deployment/7-disable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:37:35 | sidecar-deployment/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:37:35 | sidecar-deployment | sidecar-deployment events from ns kuttl-test-becoming-kite: logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:46 +0000 UTC Normal Pod agent-as-sidecar-6fd45d4596-jrwnv Binding Scheduled Successfully assigned kuttl-test-becoming-kite/agent-as-sidecar-6fd45d4596-jrwnv to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:46 +0000 UTC Normal Pod agent-as-sidecar-6fd45d4596-jrwnv AddedInterface Add eth0 [10.131.0.56/23] from ovn-kubernetes multus logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:46 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-6fd45d4596 SuccessfulCreate Created pod: agent-as-sidecar-6fd45d4596-jrwnv replicaset-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:46 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-6fd45d4596 to 1 deployment-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:47 +0000 UTC Normal Pod agent-as-sidecar-6fd45d4596-jrwnv.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:47 +0000 UTC Normal Pod agent-as-sidecar-6fd45d4596-jrwnv.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:47 +0000 UTC Normal Pod agent-as-sidecar-6fd45d4596-jrwnv.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:48 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-rh4jl Binding Scheduled Successfully assigned kuttl-test-becoming-kite/vertx-create-span-sidecar-84d458b68c-rh4jl to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:48 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-rh4jl replicaset-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:48 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:49 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-rh4jl AddedInterface Add eth0 [10.129.2.63/23] from ovn-kubernetes multus logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:49 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-rh4jl.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:49 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-rh4jl.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:49 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-rh4jl.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:51 +0000 UTC Normal Pod vertx-create-span-sidecar-6d4c47849-56wb8 Binding Scheduled Successfully assigned kuttl-test-becoming-kite/vertx-create-span-sidecar-6d4c47849-56wb8 to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:51 +0000 UTC Normal Pod vertx-create-span-sidecar-6d4c47849-56wb8 AddedInterface Add eth0 [10.128.2.57/23] from ovn-kubernetes multus logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:51 +0000 UTC Normal Pod vertx-create-span-sidecar-6d4c47849-56wb8.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:51 +0000 UTC Normal Pod vertx-create-span-sidecar-6d4c47849-56wb8.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:51 +0000 UTC Normal Pod vertx-create-span-sidecar-6d4c47849-56wb8.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:51 +0000 UTC Normal Pod vertx-create-span-sidecar-6d4c47849-56wb8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:51 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6d4c47849 SuccessfulCreate Created pod: vertx-create-span-sidecar-6d4c47849-56wb8 replicaset-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:51 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6d4c47849 to 1 deployment-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:52 +0000 UTC Normal Pod vertx-create-span-sidecar-6d4c47849-56wb8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:52 +0000 UTC Normal Pod vertx-create-span-sidecar-6d4c47849-56wb8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:53 +0000 UTC Normal Pod 00-find-service-xc2sb Binding Scheduled Successfully assigned kuttl-test-becoming-kite/00-find-service-xc2sb to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:53 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-xc2sb job-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:54 +0000 UTC Normal Pod 00-find-service-xc2sb AddedInterface Add eth0 [10.129.2.64/23] from ovn-kubernetes multus logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:54 +0000 UTC Normal Pod 00-find-service-xc2sb.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:54 +0000 UTC Normal Pod 00-find-service-xc2sb.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 392ms (392ms including waiting) kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:54 +0000 UTC Normal Pod 00-find-service-xc2sb.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:54 +0000 UTC Normal Pod 00-find-service-xc2sb.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:57 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-rh4jl.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.63:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:57 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-rh4jl.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.63:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:59 +0000 UTC Warning Pod vertx-create-span-sidecar-6d4c47849-56wb8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.57:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:59 +0000 UTC Warning Pod vertx-create-span-sidecar-6d4c47849-56wb8.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.57:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:59 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-rh4jl.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:59 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-rh4jl.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.63:8080/": read tcp 10.129.2.2:44752->10.129.2.63:8080: read: connection reset by peer kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:36:59 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-rh4jl.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.63:8080/": dial tcp 10.129.2.63:8080: connect: connection refused kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:01 +0000 UTC Normal Pod vertx-create-span-sidecar-6d4c47849-56wb8.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:02 +0000 UTC Warning Pod vertx-create-span-sidecar-6d4c47849-56wb8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.57:8080/": read tcp 10.128.2.2:37720->10.128.2.57:8080: read: connection reset by peer kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:02 +0000 UTC Warning Pod vertx-create-span-sidecar-6d4c47849-56wb8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.57:8080/": dial tcp 10.128.2.57:8080: connect: connection refused kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:05 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:09 +0000 UTC Normal Pod agent-as-sidecar2-869f99846b-jwk8f Binding Scheduled Successfully assigned kuttl-test-becoming-kite/agent-as-sidecar2-869f99846b-jwk8f to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:09 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-869f99846b SuccessfulCreate Created pod: agent-as-sidecar2-869f99846b-jwk8f replicaset-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:09 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-869f99846b to 1 deployment-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:10 +0000 UTC Normal Pod agent-as-sidecar2-869f99846b-jwk8f AddedInterface Add eth0 [10.131.0.57/23] from ovn-kubernetes multus logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:10 +0000 UTC Normal Pod agent-as-sidecar2-869f99846b-jwk8f.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:10 +0000 UTC Normal Pod agent-as-sidecar2-869f99846b-jwk8f.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:10 +0000 UTC Normal Pod agent-as-sidecar2-869f99846b-jwk8f.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:10 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-rh4jl.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.63:8080/": read tcp 10.129.2.2:42396->10.129.2.63:8080: read: connection reset by peer kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:12 +0000 UTC Normal Pod 01-find-service-lfzzn Binding Scheduled Successfully assigned kuttl-test-becoming-kite/01-find-service-lfzzn to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:12 +0000 UTC Normal Pod 01-find-service-lfzzn AddedInterface Add eth0 [10.129.2.65/23] from ovn-kubernetes multus logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:12 +0000 UTC Normal Pod 01-find-service-lfzzn.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:12 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-lfzzn job-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:12 +0000 UTC Normal Pod agent-as-sidecar-6fd45d4596-jrwnv.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:13 +0000 UTC Normal Pod 01-find-service-lfzzn.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 389ms (389ms including waiting) kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:13 +0000 UTC Normal Pod 01-find-service-lfzzn.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:13 +0000 UTC Normal Pod 01-find-service-lfzzn.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:16 +0000 UTC Normal Pod vertx-create-span-sidecar-6fc4c975dc-dnfwn Binding Scheduled Successfully assigned kuttl-test-becoming-kite/vertx-create-span-sidecar-6fc4c975dc-dnfwn to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:16 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6fc4c975dc SuccessfulCreate Created pod: vertx-create-span-sidecar-6fc4c975dc-dnfwn replicaset-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:16 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-rh4jl replicaset-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:16 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:16 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6fc4c975dc to 1 from 0 deployment-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:17 +0000 UTC Normal Pod vertx-create-span-sidecar-6fc4c975dc-dnfwn AddedInterface Add eth0 [10.131.0.58/23] from ovn-kubernetes multus logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:17 +0000 UTC Normal Pod vertx-create-span-sidecar-6fc4c975dc-dnfwn.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:17 +0000 UTC Normal Pod vertx-create-span-sidecar-6fc4c975dc-dnfwn.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:17 +0000 UTC Normal Pod vertx-create-span-sidecar-6fc4c975dc-dnfwn.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:17 +0000 UTC Normal Pod vertx-create-span-sidecar-6fc4c975dc-dnfwn.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:17 +0000 UTC Normal Pod vertx-create-span-sidecar-6fc4c975dc-dnfwn.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:17 +0000 UTC Normal Pod vertx-create-span-sidecar-6fc4c975dc-dnfwn.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:25 +0000 UTC Warning Pod vertx-create-span-sidecar-6fc4c975dc-dnfwn.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.58:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:25 +0000 UTC Warning Pod vertx-create-span-sidecar-6fc4c975dc-dnfwn.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.58:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:28 +0000 UTC Normal Pod vertx-create-span-sidecar-6fc4c975dc-dnfwn.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:28 +0000 UTC Warning Pod vertx-create-span-sidecar-6fc4c975dc-dnfwn.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.58:8080/": read tcp 10.131.0.2:48170->10.131.0.58:8080: read: connection reset by peer kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:28 +0000 UTC Warning Pod vertx-create-span-sidecar-6fc4c975dc-dnfwn.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.58:8080/": dial tcp 10.131.0.58:8080: connect: connection refused kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:32 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:32 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6d4c47849 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-6d4c47849-56wb8 replicaset-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:32 +0000 UTC Normal Pod vertx-create-span-sidecar-7945d9b494-qs2m2 Binding Scheduled Successfully assigned kuttl-test-becoming-kite/vertx-create-span-sidecar-7945d9b494-qs2m2 to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:32 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-7945d9b494 SuccessfulCreate Created pod: vertx-create-span-sidecar-7945d9b494-qs2m2 replicaset-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:32 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-6d4c47849 to 0 from 1 deployment-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:32 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-7945d9b494 to 1 from 0 deployment-controller logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:33 +0000 UTC Normal Pod vertx-create-span-sidecar-7945d9b494-qs2m2 AddedInterface Add eth0 [10.129.2.66/23] from ovn-kubernetes multus logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:33 +0000 UTC Normal Pod vertx-create-span-sidecar-7945d9b494-qs2m2.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:33 +0000 UTC Normal Pod vertx-create-span-sidecar-7945d9b494-qs2m2.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:37:35 | sidecar-deployment | 2024-06-10 07:37:33 +0000 UTC Normal Pod vertx-create-span-sidecar-7945d9b494-qs2m2.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:37:35 | sidecar-deployment | Deleting namespace: kuttl-test-becoming-kite === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (146.98s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.93s) --- PASS: kuttl/harness/sidecar-namespace (64.19s) --- PASS: kuttl/harness/sidecar-skip-webhook (17.29s) --- PASS: kuttl/harness/sidecar-deployment (59.38s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name sidecar --report --output /logs/artifacts/sidecar.xml ./artifacts/kuttl-report.xml time="2024-06-10T07:37:42Z" level=debug msg="Setting a new name for the test suites" time="2024-06-10T07:37:42Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-06-10T07:37:42Z" level=debug msg="normalizing test case names" time="2024-06-10T07:37:42Z" level=debug msg="sidecar/artifacts -> sidecar_artifacts" time="2024-06-10T07:37:42Z" level=debug msg="sidecar/sidecar-namespace -> sidecar_sidecar_namespace" time="2024-06-10T07:37:42Z" level=debug msg="sidecar/sidecar-skip-webhook -> sidecar_sidecar_skip_webhook" time="2024-06-10T07:37:42Z" level=debug msg="sidecar/sidecar-deployment -> sidecar_sidecar_deployment" +------------------------------+--------+ | NAME | RESULT | +------------------------------+--------+ | sidecar_artifacts | passed | | sidecar_sidecar_namespace | passed | | sidecar_sidecar_skip_webhook | passed | | sidecar_sidecar_deployment | passed | +------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh streaming false true + '[' 3 -ne 3 ']' + test_suite_name=streaming + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/streaming.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-streaming make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ SKIP_ES_EXTERNAL=true \ ./tests/e2e/streaming/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-07-024651 True False 51m Cluster version is 4.16.0-0.nightly-2024-06-07-024651' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-07-024651 True False 51m Cluster version is 4.16.0-0.nightly-2024-06-07-024651' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/streaming/render.sh ++ export SUITE_DIR=./tests/e2e/streaming ++ SUITE_DIR=./tests/e2e/streaming ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/streaming ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + '[' false = true ']' + start_test streaming-simple + '[' 1 -ne 1 ']' + test_name=streaming-simple + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-simple' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-simple\e[0m' Rendering files for test streaming-simple + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + mkdir -p streaming-simple + cd streaming-simple + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/streaming-jaeger-assert.yaml.template -o ./04-assert.yaml + render_smoke_test simple-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=simple-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + export JAEGER_NAME=simple-streaming + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-tls + '[' 1 -ne 1 ']' + test_name=streaming-with-tls + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-tls' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-tls\e[0m' Rendering files for test streaming-with-tls + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-simple + '[' streaming-simple '!=' _build ']' + cd .. + mkdir -p streaming-with-tls + cd streaming-with-tls + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + render_smoke_test tls-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=tls-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + export JAEGER_NAME=tls-streaming + JAEGER_NAME=tls-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-autoprovisioning-autoscale + '[' 1 -ne 1 ']' + test_name=streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-autoprovisioning-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-autoprovisioning-autoscale\e[0m' Rendering files for test streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-tls + '[' streaming-with-tls '!=' _build ']' + cd .. + mkdir -p streaming-with-autoprovisioning-autoscale + cd streaming-with-autoprovisioning-autoscale + '[' true = true ']' + rm ./00-install.yaml ./00-assert.yaml + render_install_elasticsearch upstream 01 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=01 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./01-assert.yaml + jaeger_name=auto-provisioned + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="20Mi"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="500m"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.autoscale=true ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.minReplicas=1 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.maxReplicas=2 ./02-install.yaml + render_assert_kafka true auto-provisioned 03 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provisioned + test_step=03 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./03-assert.yaml ++ expr 03 + 1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./04-assert.yaml ++ expr 03 + 2 + CLUSTER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./05-assert.yaml + version_lt 1.27 1.23 ++ echo 1.27 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.27 '!=' 1.27 + rm ./08-assert.yaml + skip_test streaming-with-tls 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-with-tls + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-autoprovisioning-autoscale + '[' streaming-with-autoprovisioning-autoscale '!=' _build ']' + cd .. + rm -rf streaming-with-tls + warning 'streaming-with-tls: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-with-tls: This test is flaky in Prow CI\e[0m' WAR: streaming-with-tls: This test is flaky in Prow CI + skip_test streaming-simple 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-simple + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + rm -rf streaming-simple + warning 'streaming-simple: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-simple: This test is flaky in Prow CI\e[0m' WAR: streaming-simple: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running streaming E2E tests' Running streaming E2E tests + cd tests/e2e/streaming/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3357233117 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 2 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/streaming-with-autoprovisioning-autoscale === PAUSE kuttl/harness/streaming-with-autoprovisioning-autoscale === CONT kuttl/harness/artifacts logger.go:42: 07:37:54 | artifacts | Creating namespace: kuttl-test-game-dove logger.go:42: 07:37:54 | artifacts | artifacts events from ns kuttl-test-game-dove: logger.go:42: 07:37:54 | artifacts | Deleting namespace: kuttl-test-game-dove === CONT kuttl/harness/streaming-with-autoprovisioning-autoscale logger.go:42: 07:38:00 | streaming-with-autoprovisioning-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:38:00 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:38:00 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:38:00 | streaming-with-autoprovisioning-autoscale | Creating namespace: kuttl-test-nearby-snapper logger.go:42: 07:38:00 | streaming-with-autoprovisioning-autoscale/1-install | starting test step 1-install logger.go:42: 07:38:00 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc create sa deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:38:00 | streaming-with-autoprovisioning-autoscale/1-install | serviceaccount/deploy-elasticsearch created logger.go:42: 07:38:00 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc adm policy add-scc-to-user privileged -z deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:38:00 | streaming-with-autoprovisioning-autoscale/1-install | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:privileged added: "deploy-elasticsearch" logger.go:42: 07:38:00 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 6] logger.go:42: 07:38:06 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_0.yml -n $NAMESPACE] logger.go:42: 07:38:07 | streaming-with-autoprovisioning-autoscale/1-install | statefulset.apps/elasticsearch created logger.go:42: 07:38:07 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 3] logger.go:42: 07:38:10 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_1.yml -n $NAMESPACE] logger.go:42: 07:38:10 | streaming-with-autoprovisioning-autoscale/1-install | service/elasticsearch created logger.go:42: 07:38:28 | streaming-with-autoprovisioning-autoscale/1-install | test step completed 1-install logger.go:42: 07:38:28 | streaming-with-autoprovisioning-autoscale/2-install | starting test step 2-install logger.go:42: 07:38:28 | streaming-with-autoprovisioning-autoscale/2-install | Jaeger:kuttl-test-nearby-snapper/auto-provisioned created logger.go:42: 07:38:28 | streaming-with-autoprovisioning-autoscale/2-install | test step completed 2-install logger.go:42: 07:38:28 | streaming-with-autoprovisioning-autoscale/3- | starting test step 3- logger.go:42: 07:39:16 | streaming-with-autoprovisioning-autoscale/3- | test step completed 3- logger.go:42: 07:39:16 | streaming-with-autoprovisioning-autoscale/4- | starting test step 4- logger.go:42: 07:39:46 | streaming-with-autoprovisioning-autoscale/4- | test step completed 4- logger.go:42: 07:39:46 | streaming-with-autoprovisioning-autoscale/5- | starting test step 5- logger.go:42: 07:40:09 | streaming-with-autoprovisioning-autoscale/5- | test step completed 5- logger.go:42: 07:40:09 | streaming-with-autoprovisioning-autoscale/6- | starting test step 6- logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale/6- | test step completed 6- logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale/7- | starting test step 7- logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale/7- | test step completed 7- logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | streaming-with-autoprovisioning-autoscale events from ns kuttl-test-nearby-snapper: logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:07 +0000 UTC Normal Pod elasticsearch-0 Binding Scheduled Successfully assigned kuttl-test-nearby-snapper/elasticsearch-0 to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:07 +0000 UTC Normal Pod elasticsearch-0 AddedInterface Add eth0 [10.131.0.59/23] from ovn-kubernetes multus logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:07 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulling Pulling image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:07 +0000 UTC Normal StatefulSet.apps elasticsearch SuccessfulCreate create Pod elasticsearch-0 in StatefulSet elasticsearch successful statefulset-controller logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:15 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulled Successfully pulled image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" in 8.102s (8.102s including waiting) kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:15 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:15 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:22 +0000 UTC Warning Pod elasticsearch-0.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Get "http://10.131.0.59:9200/": dial tcp 10.131.0.59:9200: connect: connection refused kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:35 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-zookeeper NoPods No matching pods found controllermanager logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:35 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:35 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:35 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-nearby-snapper/data-auto-provisioned-zookeeper-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-5598ddb9d6-v65mx_106d1031-cf71-4741-9ed6-0e88afff9f5e logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:40 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 Binding Scheduled Successfully assigned kuttl-test-nearby-snapper/auto-provisioned-zookeeper-0 to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:40 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ProvisioningSucceeded Successfully provisioned volume pvc-08722105-d164-4c37-9ce3-6a724ff37bb6 ebs.csi.aws.com_aws-ebs-csi-driver-controller-5598ddb9d6-v65mx_106d1031-cf71-4741-9ed6-0e88afff9f5e logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:42 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-08722105-d164-4c37-9ce3-6a724ff37bb6" attachdetach-controller logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:46 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 AddedInterface Add eth0 [10.129.2.67/23] from ovn-kubernetes multus logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:46 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:6d800bc0aa810993a93517c2464eb4575ee0d2579a1b0d84f3b8614550e32c94" kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:55 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:6d800bc0aa810993a93517c2464eb4575ee0d2579a1b0d84f3b8614550e32c94" in 9.201s (9.201s including waiting) kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:55 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Created Created container zookeeper kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:38:55 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Started Started container zookeeper kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:17 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:18 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-kafka NoPods No matching pods found controllermanager logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:18 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-nearby-snapper/data-0-auto-provisioned-kafka-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-5598ddb9d6-v65mx_106d1031-cf71-4741-9ed6-0e88afff9f5e logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:18 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:22 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ProvisioningSucceeded Successfully provisioned volume pvc-daf80880-9439-4686-93ee-52c5b582258e ebs.csi.aws.com_aws-ebs-csi-driver-controller-5598ddb9d6-v65mx_106d1031-cf71-4741-9ed6-0e88afff9f5e logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:23 +0000 UTC Normal Pod auto-provisioned-kafka-0 Binding Scheduled Successfully assigned kuttl-test-nearby-snapper/auto-provisioned-kafka-0 to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:25 +0000 UTC Normal Pod auto-provisioned-kafka-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-daf80880-9439-4686-93ee-52c5b582258e" attachdetach-controller logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:26 +0000 UTC Normal Pod auto-provisioned-kafka-0 AddedInterface Add eth0 [10.128.2.58/23] from ovn-kubernetes multus logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:26 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:6d800bc0aa810993a93517c2464eb4575ee0d2579a1b0d84f3b8614550e32c94" kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:31 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:6d800bc0aa810993a93517c2464eb4575ee0d2579a1b0d84f3b8614550e32c94" in 5.256s (5.256s including waiting) kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:31 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Created Created container kafka kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:31 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Started Started container kafka kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:48 +0000 UTC Normal Pod auto-provisioned-entity-operator-d9db89fbc-q5k5n Binding Scheduled Successfully assigned kuttl-test-nearby-snapper/auto-provisioned-entity-operator-d9db89fbc-q5k5n to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:48 +0000 UTC Normal ReplicaSet.apps auto-provisioned-entity-operator-d9db89fbc SuccessfulCreate Created pod: auto-provisioned-entity-operator-d9db89fbc-q5k5n replicaset-controller logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:48 +0000 UTC Normal Deployment.apps auto-provisioned-entity-operator ScalingReplicaSet Scaled up replica set auto-provisioned-entity-operator-d9db89fbc to 1 deployment-controller logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:49 +0000 UTC Normal Pod auto-provisioned-entity-operator-d9db89fbc-q5k5n AddedInterface Add eth0 [10.128.2.59/23] from ovn-kubernetes multus logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:49 +0000 UTC Normal Pod auto-provisioned-entity-operator-d9db89fbc-q5k5n.spec.containers{topic-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:25768564d6d2b15ac00c95e595878716001d49bf896599d31b4fa6e17b64e8c4" already present on machine kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:49 +0000 UTC Normal Pod auto-provisioned-entity-operator-d9db89fbc-q5k5n.spec.containers{topic-operator} Created Created container topic-operator kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:49 +0000 UTC Normal Pod auto-provisioned-entity-operator-d9db89fbc-q5k5n.spec.containers{topic-operator} Started Started container topic-operator kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:49 +0000 UTC Normal Pod auto-provisioned-entity-operator-d9db89fbc-q5k5n.spec.containers{user-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:25768564d6d2b15ac00c95e595878716001d49bf896599d31b4fa6e17b64e8c4" already present on machine kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:49 +0000 UTC Normal Pod auto-provisioned-entity-operator-d9db89fbc-q5k5n.spec.containers{user-operator} Created Created container user-operator kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:39:49 +0000 UTC Normal Pod auto-provisioned-entity-operator-d9db89fbc-q5k5n.spec.containers{user-operator} Started Started container user-operator kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:11 +0000 UTC Normal Pod auto-provisioned-collector-684856c69-rr8b6 Binding Scheduled Successfully assigned kuttl-test-nearby-snapper/auto-provisioned-collector-684856c69-rr8b6 to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:11 +0000 UTC Normal ReplicaSet.apps auto-provisioned-collector-684856c69 SuccessfulCreate Created pod: auto-provisioned-collector-684856c69-rr8b6 replicaset-controller logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:11 +0000 UTC Normal Deployment.apps auto-provisioned-collector ScalingReplicaSet Scaled up replica set auto-provisioned-collector-684856c69 to 1 deployment-controller logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:11 +0000 UTC Normal Pod auto-provisioned-ingester-9686757c8-qxldg Binding Scheduled Successfully assigned kuttl-test-nearby-snapper/auto-provisioned-ingester-9686757c8-qxldg to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:11 +0000 UTC Normal ReplicaSet.apps auto-provisioned-ingester-9686757c8 SuccessfulCreate Created pod: auto-provisioned-ingester-9686757c8-qxldg replicaset-controller logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:11 +0000 UTC Normal Deployment.apps auto-provisioned-ingester ScalingReplicaSet Scaled up replica set auto-provisioned-ingester-9686757c8 to 1 deployment-controller logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:11 +0000 UTC Normal Pod auto-provisioned-query-566465b9d4-48kgc Binding Scheduled Successfully assigned kuttl-test-nearby-snapper/auto-provisioned-query-566465b9d4-48kgc to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:11 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-566465b9d4 SuccessfulCreate Created pod: auto-provisioned-query-566465b9d4-48kgc replicaset-controller logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:11 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled up replica set auto-provisioned-query-566465b9d4 to 1 deployment-controller logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:12 +0000 UTC Normal Pod auto-provisioned-collector-684856c69-rr8b6 AddedInterface Add eth0 [10.131.0.60/23] from ovn-kubernetes multus logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:12 +0000 UTC Normal Pod auto-provisioned-collector-684856c69-rr8b6.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" already present on machine kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:12 +0000 UTC Normal Pod auto-provisioned-collector-684856c69-rr8b6.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:12 +0000 UTC Normal Pod auto-provisioned-collector-684856c69-rr8b6.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:12 +0000 UTC Normal Pod auto-provisioned-ingester-9686757c8-qxldg AddedInterface Add eth0 [10.129.2.68/23] from ovn-kubernetes multus logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:12 +0000 UTC Normal Pod auto-provisioned-ingester-9686757c8-qxldg.spec.containers{jaeger-ingester} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:d6c1f6133afac3376b8318c5f67e78b6f0434ec271ab7b4890e4c784af61cfb6" kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:12 +0000 UTC Normal Pod auto-provisioned-query-566465b9d4-48kgc AddedInterface Add eth0 [10.129.2.69/23] from ovn-kubernetes multus logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:12 +0000 UTC Normal Pod auto-provisioned-query-566465b9d4-48kgc.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:12 +0000 UTC Normal Pod auto-provisioned-query-566465b9d4-48kgc.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:12 +0000 UTC Normal Pod auto-provisioned-query-566465b9d4-48kgc.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:12 +0000 UTC Normal Pod auto-provisioned-query-566465b9d4-48kgc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:12 +0000 UTC Normal Pod auto-provisioned-query-566465b9d4-48kgc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:12 +0000 UTC Normal Pod auto-provisioned-query-566465b9d4-48kgc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:12 +0000 UTC Normal Pod auto-provisioned-query-566465b9d4-48kgc.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:12 +0000 UTC Normal Pod auto-provisioned-query-566465b9d4-48kgc.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:12 +0000 UTC Normal Pod auto-provisioned-query-566465b9d4-48kgc.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:16 +0000 UTC Normal Pod auto-provisioned-ingester-9686757c8-qxldg.spec.containers{jaeger-ingester} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:d6c1f6133afac3376b8318c5f67e78b6f0434ec271ab7b4890e4c784af61cfb6" in 4.195s (4.195s including waiting) kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:16 +0000 UTC Normal Pod auto-provisioned-ingester-9686757c8-qxldg.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:16 +0000 UTC Normal Pod auto-provisioned-ingester-9686757c8-qxldg.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | 2024-06-10 07:40:17 +0000 UTC Warning Pod auto-provisioned-ingester-9686757c8-qxldg.spec.containers{jaeger-ingester} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 07:40:19 | streaming-with-autoprovisioning-autoscale | Deleting namespace: kuttl-test-nearby-snapper === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (185.54s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.94s) --- PASS: kuttl/harness/streaming-with-autoprovisioning-autoscale (179.41s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name streaming --report --output /logs/artifacts/streaming.xml ./artifacts/kuttl-report.xml time="2024-06-10T07:40:59Z" level=debug msg="Setting a new name for the test suites" time="2024-06-10T07:40:59Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-06-10T07:40:59Z" level=debug msg="normalizing test case names" time="2024-06-10T07:40:59Z" level=debug msg="streaming/artifacts -> streaming_artifacts" time="2024-06-10T07:40:59Z" level=debug msg="streaming/streaming-with-autoprovisioning-autoscale -> streaming_streaming_with_autoprovisioning_autoscale" +-----------------------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------------------+--------+ | streaming_artifacts | passed | | streaming_streaming_with_autoprovisioning_autoscale | passed | +-----------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh ui false true + '[' 3 -ne 3 ']' + test_suite_name=ui + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/ui.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-ui make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true ./tests/e2e/ui/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-07-024651 True False 54m Cluster version is 4.16.0-0.nightly-2024-06-07-024651' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-07-024651 True False 54m Cluster version is 4.16.0-0.nightly-2024-06-07-024651' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/ui/render.sh ++ export SUITE_DIR=./tests/e2e/ui ++ SUITE_DIR=./tests/e2e/ui ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/ui ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test allinone + '[' 1 -ne 1 ']' + test_name=allinone + echo =========================================================================== =========================================================================== + info 'Rendering files for test allinone' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test allinone\e[0m' Rendering files for test allinone + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build + '[' _build '!=' _build ']' + mkdir -p allinone + cd allinone + export GET_URL_COMMAND + export URL + export JAEGER_NAME=all-in-one-ui + JAEGER_NAME=all-in-one-ui + '[' true = true ']' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./01-curl.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./04-test-ui-config.yaml + start_test production + '[' 1 -ne 1 ']' + test_name=production + echo =========================================================================== =========================================================================== + info 'Rendering files for test production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test production\e[0m' Rendering files for test production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build/allinone + '[' allinone '!=' _build ']' + cd .. + mkdir -p production + cd production + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + [[ true = true ]] + [[ true = true ]] + render_install_jaeger production-ui production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + '[' true = true ']' + INSECURE=true + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-forbbiden-access.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-curl.yaml + INSECURE=true + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./05-check-disabled-security.yaml + ASSERT_PRESENT=false + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./06-check-NO-gaID.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./08-check-gaID.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running ui E2E tests' Running ui E2E tests + cd tests/e2e/ui/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3357233117 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 3 tests === RUN kuttl/harness === RUN kuttl/harness/allinone === PAUSE kuttl/harness/allinone === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/production === PAUSE kuttl/harness/production === CONT kuttl/harness/allinone logger.go:42: 07:41:06 | allinone | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:41:06 | allinone | Creating namespace: kuttl-test-beloved-bluebird logger.go:42: 07:41:06 | allinone/0-install | starting test step 0-install logger.go:42: 07:41:06 | allinone/0-install | Jaeger:kuttl-test-beloved-bluebird/all-in-one-ui created logger.go:42: 07:41:10 | allinone/0-install | test step completed 0-install logger.go:42: 07:41:10 | allinone/1-curl | starting test step 1-curl logger.go:42: 07:41:10 | allinone/1-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 07:41:10 | allinone/1-curl | Checking the Ingress host value was populated logger.go:42: 07:41:10 | allinone/1-curl | Try number 0 logger.go:42: 07:41:10 | allinone/1-curl | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 07:41:10 | allinone/1-curl | template was: logger.go:42: 07:41:10 | allinone/1-curl | {.items[0].status.ingress[0].host} logger.go:42: 07:41:10 | allinone/1-curl | object given to jsonpath engine was: logger.go:42: 07:41:10 | allinone/1-curl | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 07:41:10 | allinone/1-curl | logger.go:42: 07:41:10 | allinone/1-curl | logger.go:42: 07:41:20 | allinone/1-curl | Try number 1 logger.go:42: 07:41:20 | allinone/1-curl | Hostname is all-in-one-ui-kuttl-test-beloved-bluebird.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com logger.go:42: 07:41:20 | allinone/1-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE all-in-one-ui] logger.go:42: 07:41:21 | allinone/1-curl | Checking an expected HTTP response logger.go:42: 07:41:21 | allinone/1-curl | Running in OpenShift logger.go:42: 07:41:21 | allinone/1-curl | User not provided. Getting the token... logger.go:42: 07:41:22 | allinone/1-curl | Warning: resource jaegers/all-in-one-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:41:28 | allinone/1-curl | Try number 1/30 the https://all-in-one-ui-kuttl-test-beloved-bluebird.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 07:41:28 | allinone/1-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:41:28 | allinone/1-curl | Try number 2/30 the https://all-in-one-ui-kuttl-test-beloved-bluebird.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 07:41:29 | allinone/1-curl | curl response asserted properly logger.go:42: 07:41:29 | allinone/1-curl | test step completed 1-curl logger.go:42: 07:41:29 | allinone/2-delete | starting test step 2-delete logger.go:42: 07:41:29 | allinone/2-delete | Jaeger:kuttl-test-beloved-bluebird/all-in-one-ui created logger.go:42: 07:41:29 | allinone/2-delete | test step completed 2-delete logger.go:42: 07:41:29 | allinone/3-install | starting test step 3-install logger.go:42: 07:41:29 | allinone/3-install | Jaeger:kuttl-test-beloved-bluebird/all-in-one-ui updated logger.go:42: 07:41:29 | allinone/3-install | test step completed 3-install logger.go:42: 07:41:29 | allinone/4-test-ui-config | starting test step 4-test-ui-config logger.go:42: 07:41:29 | allinone/4-test-ui-config | running command: [./ensure-ingress-host.sh] logger.go:42: 07:41:29 | allinone/4-test-ui-config | Checking the Ingress host value was populated logger.go:42: 07:41:29 | allinone/4-test-ui-config | Try number 0 logger.go:42: 07:41:29 | allinone/4-test-ui-config | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 07:41:29 | allinone/4-test-ui-config | template was: logger.go:42: 07:41:29 | allinone/4-test-ui-config | {.items[0].status.ingress[0].host} logger.go:42: 07:41:29 | allinone/4-test-ui-config | object given to jsonpath engine was: logger.go:42: 07:41:29 | allinone/4-test-ui-config | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 07:41:29 | allinone/4-test-ui-config | logger.go:42: 07:41:29 | allinone/4-test-ui-config | logger.go:42: 07:41:39 | allinone/4-test-ui-config | Try number 1 logger.go:42: 07:41:39 | allinone/4-test-ui-config | Hostname is all-in-one-ui-kuttl-test-beloved-bluebird.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com logger.go:42: 07:41:39 | allinone/4-test-ui-config | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:41:40 | allinone/4-test-ui-config | time="2024-06-10T07:41:40Z" level=info msg="Querying https://all-in-one-ui-kuttl-test-beloved-bluebird.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:41:40 | allinone/4-test-ui-config | time="2024-06-10T07:41:40Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:41:40 | allinone/4-test-ui-config | time="2024-06-10T07:41:40Z" level=info msg="Polling to https://all-in-one-ui-kuttl-test-beloved-bluebird.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:41:40 | allinone/4-test-ui-config | time="2024-06-10T07:41:40Z" level=info msg="Doing request number 0" logger.go:42: 07:41:40 | allinone/4-test-ui-config | time="2024-06-10T07:41:40Z" level=info msg="Content found and asserted!" logger.go:42: 07:41:40 | allinone/4-test-ui-config | time="2024-06-10T07:41:40Z" level=info msg="Success!" logger.go:42: 07:41:40 | allinone/4-test-ui-config | test step completed 4-test-ui-config logger.go:42: 07:41:40 | allinone | allinone events from ns kuttl-test-beloved-bluebird: logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:10 +0000 UTC Normal Pod all-in-one-ui-5db6fcc479-2nlrh Binding Scheduled Successfully assigned kuttl-test-beloved-bluebird/all-in-one-ui-5db6fcc479-2nlrh to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:10 +0000 UTC Normal Pod all-in-one-ui-5db6fcc479-2nlrh AddedInterface Add eth0 [10.131.0.61/23] from ovn-kubernetes multus logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:10 +0000 UTC Normal Pod all-in-one-ui-5db6fcc479-2nlrh.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:10 +0000 UTC Normal Pod all-in-one-ui-5db6fcc479-2nlrh.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:10 +0000 UTC Normal Pod all-in-one-ui-5db6fcc479-2nlrh.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:10 +0000 UTC Normal Pod all-in-one-ui-5db6fcc479-2nlrh.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:10 +0000 UTC Normal Pod all-in-one-ui-5db6fcc479-2nlrh.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:10 +0000 UTC Normal Pod all-in-one-ui-5db6fcc479-2nlrh.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:10 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-5db6fcc479 SuccessfulCreate Created pod: all-in-one-ui-5db6fcc479-2nlrh replicaset-controller logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:10 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-5db6fcc479 to 1 deployment-controller logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:23 +0000 UTC Normal Pod all-in-one-ui-5db6fcc479-2nlrh.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:23 +0000 UTC Normal Pod all-in-one-ui-5db6fcc479-2nlrh.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:23 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-5db6fcc479 SuccessfulDelete Deleted pod: all-in-one-ui-5db6fcc479-2nlrh replicaset-controller logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:23 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-5db6fcc479 to 0 from 1 deployment-controller logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:24 +0000 UTC Normal Pod all-in-one-ui-6f787cf5c9-gtn6h Binding Scheduled Successfully assigned kuttl-test-beloved-bluebird/all-in-one-ui-6f787cf5c9-gtn6h to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:24 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-6f787cf5c9 SuccessfulCreate Created pod: all-in-one-ui-6f787cf5c9-gtn6h replicaset-controller logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:24 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-6f787cf5c9 to 1 deployment-controller logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:25 +0000 UTC Normal Pod all-in-one-ui-6f787cf5c9-gtn6h AddedInterface Add eth0 [10.131.0.62/23] from ovn-kubernetes multus logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:25 +0000 UTC Normal Pod all-in-one-ui-6f787cf5c9-gtn6h.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:25 +0000 UTC Normal Pod all-in-one-ui-6f787cf5c9-gtn6h.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:25 +0000 UTC Normal Pod all-in-one-ui-6f787cf5c9-gtn6h.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:25 +0000 UTC Normal Pod all-in-one-ui-6f787cf5c9-gtn6h.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:25 +0000 UTC Normal Pod all-in-one-ui-6f787cf5c9-gtn6h.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:25 +0000 UTC Normal Pod all-in-one-ui-6f787cf5c9-gtn6h.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:29 +0000 UTC Normal Pod all-in-one-ui-6f787cf5c9-gtn6h.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:29 +0000 UTC Normal Pod all-in-one-ui-6f787cf5c9-gtn6h.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:35 +0000 UTC Normal Pod all-in-one-ui-59b8d8dbf8-45tck Binding Scheduled Successfully assigned kuttl-test-beloved-bluebird/all-in-one-ui-59b8d8dbf8-45tck to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:35 +0000 UTC Normal Pod all-in-one-ui-59b8d8dbf8-45tck AddedInterface Add eth0 [10.129.2.70/23] from ovn-kubernetes multus logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:35 +0000 UTC Normal Pod all-in-one-ui-59b8d8dbf8-45tck.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:35 +0000 UTC Normal Pod all-in-one-ui-59b8d8dbf8-45tck.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:35 +0000 UTC Normal Pod all-in-one-ui-59b8d8dbf8-45tck.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:35 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-59b8d8dbf8 SuccessfulCreate Created pod: all-in-one-ui-59b8d8dbf8-45tck replicaset-controller logger.go:42: 07:41:40 | allinone | 2024-06-10 07:41:35 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-59b8d8dbf8 to 1 deployment-controller logger.go:42: 07:41:40 | allinone | Deleting namespace: kuttl-test-beloved-bluebird === CONT kuttl/harness/production logger.go:42: 07:41:46 | production | Ignoring add-tracking-id.yaml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:41:46 | production | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:41:46 | production | Creating namespace: kuttl-test-prompt-sheepdog logger.go:42: 07:41:46 | production/1-install | starting test step 1-install logger.go:42: 07:41:46 | production/1-install | Jaeger:kuttl-test-prompt-sheepdog/production-ui created logger.go:42: 07:42:23 | production/1-install | test step completed 1-install logger.go:42: 07:42:23 | production/2-check-forbbiden-access | starting test step 2-check-forbbiden-access logger.go:42: 07:42:23 | production/2-check-forbbiden-access | running command: [./ensure-ingress-host.sh] logger.go:42: 07:42:23 | production/2-check-forbbiden-access | Checking the Ingress host value was populated logger.go:42: 07:42:23 | production/2-check-forbbiden-access | Try number 0 logger.go:42: 07:42:23 | production/2-check-forbbiden-access | Hostname is production-ui-kuttl-test-prompt-sheepdog.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com logger.go:42: 07:42:23 | production/2-check-forbbiden-access | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE production-ui] logger.go:42: 07:42:24 | production/2-check-forbbiden-access | Checking an expected HTTP response logger.go:42: 07:42:24 | production/2-check-forbbiden-access | Running in OpenShift logger.go:42: 07:42:24 | production/2-check-forbbiden-access | Not using any secret logger.go:42: 07:42:24 | production/2-check-forbbiden-access | Try number 1/30 the https://production-ui-kuttl-test-prompt-sheepdog.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 07:42:24 | production/2-check-forbbiden-access | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:42:24 | production/2-check-forbbiden-access | Try number 2/30 the https://production-ui-kuttl-test-prompt-sheepdog.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 07:42:24 | production/2-check-forbbiden-access | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 07:42:34 | production/2-check-forbbiden-access | Try number 3/30 the https://production-ui-kuttl-test-prompt-sheepdog.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 07:42:34 | production/2-check-forbbiden-access | curl response asserted properly logger.go:42: 07:42:34 | production/2-check-forbbiden-access | test step completed 2-check-forbbiden-access logger.go:42: 07:42:34 | production/3-curl | starting test step 3-curl logger.go:42: 07:42:34 | production/3-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 07:42:34 | production/3-curl | Checking the Ingress host value was populated logger.go:42: 07:42:34 | production/3-curl | Try number 0 logger.go:42: 07:42:34 | production/3-curl | Hostname is production-ui-kuttl-test-prompt-sheepdog.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com logger.go:42: 07:42:34 | production/3-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 07:42:34 | production/3-curl | Checking an expected HTTP response logger.go:42: 07:42:34 | production/3-curl | Running in OpenShift logger.go:42: 07:42:34 | production/3-curl | User not provided. Getting the token... logger.go:42: 07:42:36 | production/3-curl | Warning: resource jaegers/production-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:42:42 | production/3-curl | Try number 1/30 the https://production-ui-kuttl-test-prompt-sheepdog.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 07:42:42 | production/3-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:42:42 | production/3-curl | Try number 2/30 the https://production-ui-kuttl-test-prompt-sheepdog.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 07:42:42 | production/3-curl | curl response asserted properly logger.go:42: 07:42:42 | production/3-curl | test step completed 3-curl logger.go:42: 07:42:42 | production/4-install | starting test step 4-install logger.go:42: 07:42:42 | production/4-install | Jaeger:kuttl-test-prompt-sheepdog/production-ui updated logger.go:42: 07:42:42 | production/4-install | test step completed 4-install logger.go:42: 07:42:42 | production/5-check-disabled-security | starting test step 5-check-disabled-security logger.go:42: 07:42:42 | production/5-check-disabled-security | running command: [./ensure-ingress-host.sh] logger.go:42: 07:42:42 | production/5-check-disabled-security | Checking the Ingress host value was populated logger.go:42: 07:42:42 | production/5-check-disabled-security | Try number 0 logger.go:42: 07:42:42 | production/5-check-disabled-security | Hostname is production-ui-kuttl-test-prompt-sheepdog.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com logger.go:42: 07:42:42 | production/5-check-disabled-security | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 07:42:43 | production/5-check-disabled-security | Checking an expected HTTP response logger.go:42: 07:42:43 | production/5-check-disabled-security | Running in OpenShift logger.go:42: 07:42:43 | production/5-check-disabled-security | Not using any secret logger.go:42: 07:42:43 | production/5-check-disabled-security | Try number 1/30 the https://production-ui-kuttl-test-prompt-sheepdog.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 07:42:43 | production/5-check-disabled-security | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:42:43 | production/5-check-disabled-security | Try number 2/30 the https://production-ui-kuttl-test-prompt-sheepdog.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 07:42:43 | production/5-check-disabled-security | HTTP response is 403. 200 expected. Waiting 10 s logger.go:42: 07:42:53 | production/5-check-disabled-security | Try number 3/30 the https://production-ui-kuttl-test-prompt-sheepdog.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 07:42:53 | production/5-check-disabled-security | curl response asserted properly logger.go:42: 07:42:53 | production/5-check-disabled-security | test step completed 5-check-disabled-security logger.go:42: 07:42:53 | production/6-check-NO-gaID | starting test step 6-check-NO-gaID logger.go:42: 07:42:53 | production/6-check-NO-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 07:42:53 | production/6-check-NO-gaID | Checking the Ingress host value was populated logger.go:42: 07:42:53 | production/6-check-NO-gaID | Try number 0 logger.go:42: 07:42:53 | production/6-check-NO-gaID | Hostname is production-ui-kuttl-test-prompt-sheepdog.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com logger.go:42: 07:42:53 | production/6-check-NO-gaID | running command: [sh -c ASSERT_PRESENT=false EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:42:53 | production/6-check-NO-gaID | time="2024-06-10T07:42:53Z" level=info msg="Querying https://production-ui-kuttl-test-prompt-sheepdog.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:42:53 | production/6-check-NO-gaID | time="2024-06-10T07:42:53Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:42:53 | production/6-check-NO-gaID | time="2024-06-10T07:42:53Z" level=info msg="Polling to https://production-ui-kuttl-test-prompt-sheepdog.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:42:53 | production/6-check-NO-gaID | time="2024-06-10T07:42:53Z" level=info msg="Doing request number 0" logger.go:42: 07:42:53 | production/6-check-NO-gaID | time="2024-06-10T07:42:53Z" level=info msg="Content not found and asserted it was not found!" logger.go:42: 07:42:53 | production/6-check-NO-gaID | time="2024-06-10T07:42:53Z" level=info msg="Success!" logger.go:42: 07:42:53 | production/6-check-NO-gaID | test step completed 6-check-NO-gaID logger.go:42: 07:42:53 | production/7-add-tracking-id | starting test step 7-add-tracking-id logger.go:42: 07:42:53 | production/7-add-tracking-id | running command: [sh -c kubectl apply -f add-tracking-id.yaml -n $NAMESPACE] logger.go:42: 07:42:54 | production/7-add-tracking-id | jaeger.jaegertracing.io/production-ui configured logger.go:42: 07:42:54 | production/7-add-tracking-id | test step completed 7-add-tracking-id logger.go:42: 07:42:54 | production/8-check-gaID | starting test step 8-check-gaID logger.go:42: 07:42:54 | production/8-check-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 07:42:54 | production/8-check-gaID | Checking the Ingress host value was populated logger.go:42: 07:42:54 | production/8-check-gaID | Try number 0 logger.go:42: 07:42:54 | production/8-check-gaID | Hostname is production-ui-kuttl-test-prompt-sheepdog.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com logger.go:42: 07:42:54 | production/8-check-gaID | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:42:54 | production/8-check-gaID | time="2024-06-10T07:42:54Z" level=info msg="Querying https://production-ui-kuttl-test-prompt-sheepdog.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:42:54 | production/8-check-gaID | time="2024-06-10T07:42:54Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:42:54 | production/8-check-gaID | time="2024-06-10T07:42:54Z" level=info msg="Polling to https://production-ui-kuttl-test-prompt-sheepdog.apps.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:42:54 | production/8-check-gaID | time="2024-06-10T07:42:54Z" level=info msg="Doing request number 0" logger.go:42: 07:42:54 | production/8-check-gaID | time="2024-06-10T07:42:54Z" level=warning msg="Found: false . Assert: true" logger.go:42: 07:42:54 | production/8-check-gaID | time="2024-06-10T07:42:54Z" level=warning msg="The condition of the test function was not accomplished" logger.go:42: 07:43:02 | production/8-check-gaID | time="2024-06-10T07:43:02Z" level=info msg="Doing request number 1" logger.go:42: 07:43:23 | production/8-check-gaID | time="2024-06-10T07:43:23Z" level=info msg="Content found and asserted!" logger.go:42: 07:43:23 | production/8-check-gaID | time="2024-06-10T07:43:23Z" level=info msg="Success!" logger.go:42: 07:43:23 | production/8-check-gaID | test step completed 8-check-gaID logger.go:42: 07:43:23 | production | production events from ns kuttl-test-prompt-sheepdog: logger.go:42: 07:43:23 | production | 2024-06-10 07:41:53 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestpromptsheepdogproductionui-1-7d6b8f5fc6 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestpromptsheepdogproductionui-1-7dh9mmw replicaset-controller logger.go:42: 07:43:23 | production | 2024-06-10 07:41:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptsheepdogproductionui-1-7dh9mmw Binding Scheduled Successfully assigned kuttl-test-prompt-sheepdog/elasticsearch-cdm-kuttltestpromptsheepdogproductionui-1-7dh9mmw to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:23 | production | 2024-06-10 07:41:53 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestpromptsheepdogproductionui-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestpromptsheepdogproductionui-1-7d6b8f5fc6 to 1 deployment-controller logger.go:42: 07:43:23 | production | 2024-06-10 07:41:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptsheepdogproductionui-1-7dh9mmw AddedInterface Add eth0 [10.131.0.63/23] from ovn-kubernetes multus logger.go:42: 07:43:23 | production | 2024-06-10 07:41:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptsheepdogproductionui-1-7dh9mmw.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:41:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptsheepdogproductionui-1-7dh9mmw.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:41:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptsheepdogproductionui-1-7dh9mmw.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:41:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptsheepdogproductionui-1-7dh9mmw.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:41:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptsheepdogproductionui-1-7dh9mmw.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:41:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpromptsheepdogproductionui-1-7dh9mmw.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:04 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpromptsheepdogproductionui-1-7dh9mmw.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:09 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpromptsheepdogproductionui-1-7dh9mmw.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:20 +0000 UTC Normal Pod production-ui-collector-7cdffcbf67-5p76x Binding Scheduled Successfully assigned kuttl-test-prompt-sheepdog/production-ui-collector-7cdffcbf67-5p76x to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:23 | production | 2024-06-10 07:42:20 +0000 UTC Normal ReplicaSet.apps production-ui-collector-7cdffcbf67 SuccessfulCreate Created pod: production-ui-collector-7cdffcbf67-5p76x replicaset-controller logger.go:42: 07:43:23 | production | 2024-06-10 07:42:20 +0000 UTC Normal Deployment.apps production-ui-collector ScalingReplicaSet Scaled up replica set production-ui-collector-7cdffcbf67 to 1 deployment-controller logger.go:42: 07:43:23 | production | 2024-06-10 07:42:20 +0000 UTC Normal Pod production-ui-query-6946984d49-8zxm4 Binding Scheduled Successfully assigned kuttl-test-prompt-sheepdog/production-ui-query-6946984d49-8zxm4 to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:23 | production | 2024-06-10 07:42:20 +0000 UTC Normal ReplicaSet.apps production-ui-query-6946984d49 SuccessfulCreate Created pod: production-ui-query-6946984d49-8zxm4 replicaset-controller logger.go:42: 07:43:23 | production | 2024-06-10 07:42:20 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-6946984d49 to 1 deployment-controller logger.go:42: 07:43:23 | production | 2024-06-10 07:42:21 +0000 UTC Normal Pod production-ui-collector-7cdffcbf67-5p76x AddedInterface Add eth0 [10.128.2.60/23] from ovn-kubernetes multus logger.go:42: 07:43:23 | production | 2024-06-10 07:42:21 +0000 UTC Normal Pod production-ui-collector-7cdffcbf67-5p76x.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" already present on machine kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:21 +0000 UTC Normal Pod production-ui-collector-7cdffcbf67-5p76x.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:21 +0000 UTC Normal Pod production-ui-collector-7cdffcbf67-5p76x.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:21 +0000 UTC Normal Pod production-ui-query-6946984d49-8zxm4 AddedInterface Add eth0 [10.129.2.71/23] from ovn-kubernetes multus logger.go:42: 07:43:23 | production | 2024-06-10 07:42:21 +0000 UTC Normal Pod production-ui-query-6946984d49-8zxm4.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:21 +0000 UTC Normal Pod production-ui-query-6946984d49-8zxm4.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:21 +0000 UTC Normal Pod production-ui-query-6946984d49-8zxm4.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:21 +0000 UTC Normal Pod production-ui-query-6946984d49-8zxm4.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:21 +0000 UTC Normal Pod production-ui-query-6946984d49-8zxm4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:21 +0000 UTC Normal Pod production-ui-query-6946984d49-8zxm4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:21 +0000 UTC Normal Pod production-ui-query-6946984d49-8zxm4.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:21 +0000 UTC Normal Pod production-ui-query-6946984d49-8zxm4.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:21 +0000 UTC Normal Pod production-ui-query-6946984d49-8zxm4.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:35 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:43:23 | production | 2024-06-10 07:42:35 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:43:23 | production | 2024-06-10 07:42:35 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:43:23 | production | 2024-06-10 07:42:37 +0000 UTC Normal Pod production-ui-query-6946984d49-8zxm4.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:37 +0000 UTC Normal Pod production-ui-query-6946984d49-8zxm4.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:37 +0000 UTC Normal Pod production-ui-query-6946984d49-8zxm4.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:37 +0000 UTC Normal ReplicaSet.apps production-ui-query-6946984d49 SuccessfulDelete Deleted pod: production-ui-query-6946984d49-8zxm4 replicaset-controller logger.go:42: 07:43:23 | production | 2024-06-10 07:42:37 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-6946984d49 to 0 from 1 deployment-controller logger.go:42: 07:43:23 | production | 2024-06-10 07:42:38 +0000 UTC Normal Pod production-ui-query-5c6cd88845-m48lj Binding Scheduled Successfully assigned kuttl-test-prompt-sheepdog/production-ui-query-5c6cd88845-m48lj to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:23 | production | 2024-06-10 07:42:38 +0000 UTC Normal Pod production-ui-query-5c6cd88845-m48lj AddedInterface Add eth0 [10.129.2.72/23] from ovn-kubernetes multus logger.go:42: 07:43:23 | production | 2024-06-10 07:42:38 +0000 UTC Normal Pod production-ui-query-5c6cd88845-m48lj.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:38 +0000 UTC Normal Pod production-ui-query-5c6cd88845-m48lj.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:38 +0000 UTC Normal Pod production-ui-query-5c6cd88845-m48lj.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:38 +0000 UTC Normal Pod production-ui-query-5c6cd88845-m48lj.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:38 +0000 UTC Normal Pod production-ui-query-5c6cd88845-m48lj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:38 +0000 UTC Normal ReplicaSet.apps production-ui-query-5c6cd88845 SuccessfulCreate Created pod: production-ui-query-5c6cd88845-m48lj replicaset-controller logger.go:42: 07:43:23 | production | 2024-06-10 07:42:38 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-5c6cd88845 to 1 deployment-controller logger.go:42: 07:43:23 | production | 2024-06-10 07:42:39 +0000 UTC Normal Pod production-ui-query-5c6cd88845-m48lj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:39 +0000 UTC Normal Pod production-ui-query-5c6cd88845-m48lj.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:39 +0000 UTC Normal Pod production-ui-query-5c6cd88845-m48lj.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:39 +0000 UTC Normal Pod production-ui-query-5c6cd88845-m48lj.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:44 +0000 UTC Normal Pod production-ui-query-5c6cd88845-m48lj.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:44 +0000 UTC Normal Pod production-ui-query-5c6cd88845-m48lj.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:44 +0000 UTC Normal Pod production-ui-query-5c6cd88845-m48lj.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:44 +0000 UTC Normal ReplicaSet.apps production-ui-query-5c6cd88845 SuccessfulDelete Deleted pod: production-ui-query-5c6cd88845-m48lj replicaset-controller logger.go:42: 07:43:23 | production | 2024-06-10 07:42:44 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-5c6cd88845 to 0 from 1 deployment-controller logger.go:42: 07:43:23 | production | 2024-06-10 07:42:45 +0000 UTC Normal Pod production-ui-query-84487b578b-2h66d Binding Scheduled Successfully assigned kuttl-test-prompt-sheepdog/production-ui-query-84487b578b-2h66d to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:23 | production | 2024-06-10 07:42:45 +0000 UTC Normal Pod production-ui-query-84487b578b-2h66d AddedInterface Add eth0 [10.129.2.73/23] from ovn-kubernetes multus logger.go:42: 07:43:23 | production | 2024-06-10 07:42:45 +0000 UTC Normal Pod production-ui-query-84487b578b-2h66d.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:45 +0000 UTC Normal Pod production-ui-query-84487b578b-2h66d.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:45 +0000 UTC Normal Pod production-ui-query-84487b578b-2h66d.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:45 +0000 UTC Normal Pod production-ui-query-84487b578b-2h66d.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:45 +0000 UTC Normal Pod production-ui-query-84487b578b-2h66d.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:45 +0000 UTC Normal Pod production-ui-query-84487b578b-2h66d.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:45 +0000 UTC Normal ReplicaSet.apps production-ui-query-84487b578b SuccessfulCreate Created pod: production-ui-query-84487b578b-2h66d replicaset-controller logger.go:42: 07:43:23 | production | 2024-06-10 07:42:45 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-84487b578b to 1 deployment-controller logger.go:42: 07:43:23 | production | 2024-06-10 07:42:55 +0000 UTC Normal Pod production-ui-query-84487b578b-2h66d.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:55 +0000 UTC Normal Pod production-ui-query-84487b578b-2h66d.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:55 +0000 UTC Normal ReplicaSet.apps production-ui-query-84487b578b SuccessfulDelete Deleted pod: production-ui-query-84487b578b-2h66d replicaset-controller logger.go:42: 07:43:23 | production | 2024-06-10 07:42:55 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-84487b578b to 0 from 1 deployment-controller logger.go:42: 07:43:23 | production | 2024-06-10 07:42:56 +0000 UTC Normal Pod production-ui-query-5977c58b85-9f76s Binding Scheduled Successfully assigned kuttl-test-prompt-sheepdog/production-ui-query-5977c58b85-9f76s to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:23 | production | 2024-06-10 07:42:56 +0000 UTC Normal Pod production-ui-query-5977c58b85-9f76s AddedInterface Add eth0 [10.129.2.74/23] from ovn-kubernetes multus logger.go:42: 07:43:23 | production | 2024-06-10 07:42:56 +0000 UTC Normal Pod production-ui-query-5977c58b85-9f76s.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:56 +0000 UTC Normal Pod production-ui-query-5977c58b85-9f76s.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:56 +0000 UTC Normal Pod production-ui-query-5977c58b85-9f76s.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:56 +0000 UTC Normal Pod production-ui-query-5977c58b85-9f76s.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:56 +0000 UTC Normal ReplicaSet.apps production-ui-query-5977c58b85 SuccessfulCreate Created pod: production-ui-query-5977c58b85-9f76s replicaset-controller logger.go:42: 07:43:23 | production | 2024-06-10 07:42:56 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-5977c58b85 to 1 deployment-controller logger.go:42: 07:43:23 | production | 2024-06-10 07:42:57 +0000 UTC Normal Pod production-ui-query-5977c58b85-9f76s.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:42:57 +0000 UTC Normal Pod production-ui-query-5977c58b85-9f76s.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:43:23 | production | 2024-06-10 07:43:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod production-ui-collector-7cdffcbf67-5p76x horizontal-pod-autoscaler logger.go:42: 07:43:23 | production | 2024-06-10 07:43:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod production-ui-collector-7cdffcbf67-5p76x horizontal-pod-autoscaler logger.go:42: 07:43:23 | production | 2024-06-10 07:43:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod production-ui-collector-7cdffcbf67-5p76x horizontal-pod-autoscaler logger.go:42: 07:43:23 | production | Deleting namespace: kuttl-test-prompt-sheepdog === CONT kuttl/harness/artifacts logger.go:42: 07:43:29 | artifacts | Creating namespace: kuttl-test-curious-reindeer logger.go:42: 07:43:29 | artifacts | artifacts events from ns kuttl-test-curious-reindeer: logger.go:42: 07:43:29 | artifacts | Deleting namespace: kuttl-test-curious-reindeer === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (149.39s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/allinone (39.99s) --- PASS: kuttl/harness/production (103.18s) --- PASS: kuttl/harness/artifacts (6.02s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name ui --report --output /logs/artifacts/ui.xml ./artifacts/kuttl-report.xml time="2024-06-10T07:43:35Z" level=debug msg="Setting a new name for the test suites" time="2024-06-10T07:43:35Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-06-10T07:43:35Z" level=debug msg="normalizing test case names" time="2024-06-10T07:43:35Z" level=debug msg="ui/allinone -> ui_allinone" time="2024-06-10T07:43:35Z" level=debug msg="ui/production -> ui_production" time="2024-06-10T07:43:35Z" level=debug msg="ui/artifacts -> ui_artifacts" +---------------+--------+ | NAME | RESULT | +---------------+--------+ | ui_allinone | passed | | ui_production | passed | | ui_artifacts | passed | +---------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh miscellaneous false true + '[' 3 -ne 3 ']' + test_suite_name=miscellaneous + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/miscellaneous.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-miscellaneous make[2]: Entering directory '/tmp/jaeger-tests' SKIP_ES_EXTERNAL=true ./tests/e2e/miscellaneous/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-07-024651 True False 57m Cluster version is 4.16.0-0.nightly-2024-06-07-024651' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-06-07-024651 True False 57m Cluster version is 4.16.0-0.nightly-2024-06-07-024651' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/miscellaneous/render.sh ++ export SUITE_DIR=./tests/e2e/miscellaneous ++ SUITE_DIR=./tests/e2e/miscellaneous ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/miscellaneous ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test collector-autoscale + '[' 1 -ne 1 ']' + test_name=collector-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-autoscale\e[0m' Rendering files for test collector-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p collector-autoscale + cd collector-autoscale + jaeger_name=simple-prod + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + ELASTICSEARCH_NODECOUNT=1 + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.resources.requests.memory="200m"' 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.autoscale=true 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.minReplicas=1 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.maxReplicas=2 01-install.yaml + version_lt 1.27 1.23 ++ echo 1.27 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.27 '!=' 1.27 + rm ./03-assert.yaml + generate_otlp_e2e_tests http + test_protocol=http + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-http\e[0m' Rendering files for test collector-otlp-allinone-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-autoscale + '[' collector-autoscale '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-http + cd collector-otlp-allinone-http + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger http true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-http\e[0m' Rendering files for test collector-otlp-production-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-http + '[' collector-otlp-allinone-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-http + cd collector-otlp-production-http + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger http true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + generate_otlp_e2e_tests grpc + test_protocol=grpc + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-grpc\e[0m' Rendering files for test collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-http + '[' collector-otlp-production-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-grpc + cd collector-otlp-allinone-grpc + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-grpc\e[0m' Rendering files for test collector-otlp-production-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-grpc + '[' collector-otlp-allinone-grpc '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-grpc + cd collector-otlp-production-grpc + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + '[' true = true ']' + skip_test istio 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=istio + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-grpc + '[' collector-otlp-production-grpc '!=' _build ']' + cd .. + rm -rf istio + warning 'istio: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: istio: Test not supported in OpenShift\e[0m' WAR: istio: Test not supported in OpenShift + '[' true = true ']' + skip_test outside-cluster 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=outside-cluster + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf outside-cluster + warning 'outside-cluster: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: outside-cluster: Test not supported in OpenShift\e[0m' WAR: outside-cluster: Test not supported in OpenShift + start_test set-custom-img + '[' 1 -ne 1 ']' + test_name=set-custom-img + echo =========================================================================== =========================================================================== + info 'Rendering files for test set-custom-img' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test set-custom-img\e[0m' Rendering files for test set-custom-img + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p set-custom-img + cd set-custom-img + jaeger_name=my-jaeger + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.image="test"' ./02-install.yaml + '[' true = true ']' + skip_test non-cluster-wide 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=non-cluster-wide + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/set-custom-img + '[' set-custom-img '!=' _build ']' + cd .. + rm -rf non-cluster-wide + warning 'non-cluster-wide: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: non-cluster-wide: Test not supported in OpenShift\e[0m' WAR: non-cluster-wide: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running miscellaneous E2E tests' Running miscellaneous E2E tests + cd tests/e2e/miscellaneous/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3357233117 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-pvrdm7hq-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 8 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/cassandra-spark === PAUSE kuttl/harness/cassandra-spark === RUN kuttl/harness/collector-autoscale === PAUSE kuttl/harness/collector-autoscale === RUN kuttl/harness/collector-otlp-allinone-grpc === PAUSE kuttl/harness/collector-otlp-allinone-grpc === RUN kuttl/harness/collector-otlp-allinone-http === PAUSE kuttl/harness/collector-otlp-allinone-http === RUN kuttl/harness/collector-otlp-production-grpc === PAUSE kuttl/harness/collector-otlp-production-grpc === RUN kuttl/harness/collector-otlp-production-http === PAUSE kuttl/harness/collector-otlp-production-http === RUN kuttl/harness/set-custom-img === PAUSE kuttl/harness/set-custom-img === CONT kuttl/harness/artifacts logger.go:42: 07:43:46 | artifacts | Creating namespace: kuttl-test-neat-bluejay logger.go:42: 07:43:46 | artifacts | artifacts events from ns kuttl-test-neat-bluejay: logger.go:42: 07:43:46 | artifacts | Deleting namespace: kuttl-test-neat-bluejay === CONT kuttl/harness/collector-otlp-allinone-http logger.go:42: 07:43:52 | collector-otlp-allinone-http | Creating namespace: kuttl-test-cunning-koala logger.go:42: 07:43:52 | collector-otlp-allinone-http/0-install | starting test step 0-install logger.go:42: 07:43:52 | collector-otlp-allinone-http/0-install | Jaeger:kuttl-test-cunning-koala/my-jaeger created logger.go:42: 07:43:58 | collector-otlp-allinone-http/0-install | test step completed 0-install logger.go:42: 07:43:58 | collector-otlp-allinone-http/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:43:58 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:44:00 | collector-otlp-allinone-http/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:44:06 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:44:07 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:44:07 | collector-otlp-allinone-http/1-smoke-test | job.batch/report-span created logger.go:42: 07:44:07 | collector-otlp-allinone-http/1-smoke-test | job.batch/check-span created logger.go:42: 07:44:20 | collector-otlp-allinone-http/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:44:20 | collector-otlp-allinone-http | collector-otlp-allinone-http events from ns kuttl-test-cunning-koala: logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:43:56 +0000 UTC Normal Pod my-jaeger-76f556f49d-zs4m4 Binding Scheduled Successfully assigned kuttl-test-cunning-koala/my-jaeger-76f556f49d-zs4m4 to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:43:56 +0000 UTC Normal ReplicaSet.apps my-jaeger-76f556f49d SuccessfulCreate Created pod: my-jaeger-76f556f49d-zs4m4 replicaset-controller logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:43:56 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-76f556f49d to 1 deployment-controller logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:43:57 +0000 UTC Normal Pod my-jaeger-76f556f49d-zs4m4 AddedInterface Add eth0 [10.128.2.61/23] from ovn-kubernetes multus logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:43:57 +0000 UTC Normal Pod my-jaeger-76f556f49d-zs4m4.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:43:57 +0000 UTC Normal Pod my-jaeger-76f556f49d-zs4m4.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:43:57 +0000 UTC Normal Pod my-jaeger-76f556f49d-zs4m4.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:43:57 +0000 UTC Normal Pod my-jaeger-76f556f49d-zs4m4.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:43:57 +0000 UTC Normal Pod my-jaeger-76f556f49d-zs4m4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:43:57 +0000 UTC Normal Pod my-jaeger-76f556f49d-zs4m4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:03 +0000 UTC Normal Pod my-jaeger-76f556f49d-zs4m4.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:03 +0000 UTC Normal Pod my-jaeger-76f556f49d-zs4m4.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:03 +0000 UTC Normal ReplicaSet.apps my-jaeger-76f556f49d SuccessfulDelete Deleted pod: my-jaeger-76f556f49d-zs4m4 replicaset-controller logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:03 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-76f556f49d to 0 from 1 deployment-controller logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:04 +0000 UTC Normal Pod my-jaeger-68f94b98c6-4slc9 Binding Scheduled Successfully assigned kuttl-test-cunning-koala/my-jaeger-68f94b98c6-4slc9 to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:04 +0000 UTC Normal ReplicaSet.apps my-jaeger-68f94b98c6 SuccessfulCreate Created pod: my-jaeger-68f94b98c6-4slc9 replicaset-controller logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:04 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-68f94b98c6 to 1 deployment-controller logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:05 +0000 UTC Normal Pod my-jaeger-68f94b98c6-4slc9 AddedInterface Add eth0 [10.131.0.64/23] from ovn-kubernetes multus logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:05 +0000 UTC Normal Pod my-jaeger-68f94b98c6-4slc9.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:05 +0000 UTC Normal Pod my-jaeger-68f94b98c6-4slc9.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:05 +0000 UTC Normal Pod my-jaeger-68f94b98c6-4slc9.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:05 +0000 UTC Normal Pod my-jaeger-68f94b98c6-4slc9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:05 +0000 UTC Normal Pod my-jaeger-68f94b98c6-4slc9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:05 +0000 UTC Normal Pod my-jaeger-68f94b98c6-4slc9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:07 +0000 UTC Normal Pod check-span-ljprz Binding Scheduled Successfully assigned kuttl-test-cunning-koala/check-span-ljprz to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:07 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-ljprz job-controller logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:07 +0000 UTC Normal Pod report-span-zh2nj Binding Scheduled Successfully assigned kuttl-test-cunning-koala/report-span-zh2nj to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:07 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-zh2nj job-controller logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:08 +0000 UTC Normal Pod check-span-ljprz AddedInterface Add eth0 [10.128.2.62/23] from ovn-kubernetes multus logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:08 +0000 UTC Normal Pod check-span-ljprz.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:08 +0000 UTC Normal Pod check-span-ljprz.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 345ms (345ms including waiting) kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:08 +0000 UTC Normal Pod check-span-ljprz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:08 +0000 UTC Normal Pod check-span-ljprz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:08 +0000 UTC Normal Pod report-span-zh2nj AddedInterface Add eth0 [10.129.2.75/23] from ovn-kubernetes multus logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:08 +0000 UTC Normal Pod report-span-zh2nj.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:08 +0000 UTC Normal Pod report-span-zh2nj.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 365ms (365ms including waiting) kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:08 +0000 UTC Normal Pod report-span-zh2nj.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:08 +0000 UTC Normal Pod report-span-zh2nj.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:44:20 | collector-otlp-allinone-http | 2024-06-10 07:44:19 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:44:20 | collector-otlp-allinone-http | Deleting namespace: kuttl-test-cunning-koala === CONT kuttl/harness/set-custom-img logger.go:42: 07:44:31 | set-custom-img | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:44:31 | set-custom-img | Ignoring check-collector-img.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:44:31 | set-custom-img | Creating namespace: kuttl-test-sharp-airedale logger.go:42: 07:44:31 | set-custom-img/1-install | starting test step 1-install logger.go:42: 07:44:32 | set-custom-img/1-install | Jaeger:kuttl-test-sharp-airedale/my-jaeger created logger.go:42: 07:45:08 | set-custom-img/1-install | test step completed 1-install logger.go:42: 07:45:08 | set-custom-img/2-install | starting test step 2-install logger.go:42: 07:45:08 | set-custom-img/2-install | Jaeger:kuttl-test-sharp-airedale/my-jaeger updated logger.go:42: 07:45:08 | set-custom-img/2-install | test step completed 2-install logger.go:42: 07:45:08 | set-custom-img/3-check-image | starting test step 3-check-image logger.go:42: 07:45:08 | set-custom-img/3-check-image | running command: [sh -c ./check-collector-img.sh] logger.go:42: 07:45:08 | set-custom-img/3-check-image | Collector image missmatch. Expected: test. Has: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45 logger.go:42: 07:45:13 | set-custom-img/3-check-image | Collector image asserted properly! logger.go:42: 07:45:13 | set-custom-img/3-check-image | test step completed 3-check-image logger.go:42: 07:45:13 | set-custom-img | set-custom-img events from ns kuttl-test-sharp-airedale: logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:44:38 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestsharpairedalemyjaeger-1-7d6f4555dd SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestsharpairedalemyjaeger-1-7d6f455fvf67 replicaset-controller logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:44:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharpairedalemyjaeger-1-7d6f455fvf67 Binding Scheduled Successfully assigned kuttl-test-sharp-airedale/elasticsearch-cdm-kuttltestsharpairedalemyjaeger-1-7d6f455fvf67 to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:44:38 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestsharpairedalemyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestsharpairedalemyjaeger-1-7d6f4555dd to 1 deployment-controller logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:44:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharpairedalemyjaeger-1-7d6f455fvf67 AddedInterface Add eth0 [10.131.0.65/23] from ovn-kubernetes multus logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:44:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharpairedalemyjaeger-1-7d6f455fvf67.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:44:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharpairedalemyjaeger-1-7d6f455fvf67.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:44:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharpairedalemyjaeger-1-7d6f455fvf67.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:44:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharpairedalemyjaeger-1-7d6f455fvf67.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:44:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharpairedalemyjaeger-1-7d6f455fvf67.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:44:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharpairedalemyjaeger-1-7d6f455fvf67.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:44:54 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsharpairedalemyjaeger-1-7d6f455fvf67.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:05 +0000 UTC Normal Pod my-jaeger-collector-fc6c494df-mh2lv Binding Scheduled Successfully assigned kuttl-test-sharp-airedale/my-jaeger-collector-fc6c494df-mh2lv to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:05 +0000 UTC Normal Pod my-jaeger-collector-fc6c494df-mh2lv AddedInterface Add eth0 [10.128.2.64/23] from ovn-kubernetes multus logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:05 +0000 UTC Normal Pod my-jaeger-collector-fc6c494df-mh2lv.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" already present on machine kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:05 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-fc6c494df SuccessfulCreate Created pod: my-jaeger-collector-fc6c494df-mh2lv replicaset-controller logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:05 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-fc6c494df to 1 deployment-controller logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:05 +0000 UTC Normal Pod my-jaeger-query-67f7444bb4-j2pgz Binding Scheduled Successfully assigned kuttl-test-sharp-airedale/my-jaeger-query-67f7444bb4-j2pgz to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:05 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-67f7444bb4 SuccessfulCreate Created pod: my-jaeger-query-67f7444bb4-j2pgz replicaset-controller logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:05 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-67f7444bb4 to 1 deployment-controller logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:06 +0000 UTC Normal Pod my-jaeger-collector-fc6c494df-mh2lv.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:06 +0000 UTC Normal Pod my-jaeger-collector-fc6c494df-mh2lv.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:06 +0000 UTC Normal Pod my-jaeger-query-67f7444bb4-j2pgz AddedInterface Add eth0 [10.129.2.76/23] from ovn-kubernetes multus logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:06 +0000 UTC Normal Pod my-jaeger-query-67f7444bb4-j2pgz.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:06 +0000 UTC Normal Pod my-jaeger-query-67f7444bb4-j2pgz.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:06 +0000 UTC Normal Pod my-jaeger-query-67f7444bb4-j2pgz.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:06 +0000 UTC Normal Pod my-jaeger-query-67f7444bb4-j2pgz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:06 +0000 UTC Normal Pod my-jaeger-query-67f7444bb4-j2pgz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:06 +0000 UTC Normal Pod my-jaeger-query-67f7444bb4-j2pgz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:06 +0000 UTC Normal Pod my-jaeger-query-67f7444bb4-j2pgz.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:06 +0000 UTC Normal Pod my-jaeger-query-67f7444bb4-j2pgz.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:06 +0000 UTC Normal Pod my-jaeger-query-67f7444bb4-j2pgz.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:12 +0000 UTC Normal Pod my-jaeger-collector-648fbf4b47-nndkr Binding Scheduled Successfully assigned kuttl-test-sharp-airedale/my-jaeger-collector-648fbf4b47-nndkr to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:12 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-648fbf4b47 SuccessfulCreate Created pod: my-jaeger-collector-648fbf4b47-nndkr replicaset-controller logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:12 +0000 UTC Normal Pod my-jaeger-collector-fc6c494df-mh2lv.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:12 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-fc6c494df SuccessfulDelete Deleted pod: my-jaeger-collector-fc6c494df-mh2lv replicaset-controller logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:12 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-fc6c494df to 0 from 1 deployment-controller logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:12 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-648fbf4b47 to 1 deployment-controller logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:13 +0000 UTC Normal Pod my-jaeger-collector-648fbf4b47-nndkr AddedInterface Add eth0 [10.128.2.65/23] from ovn-kubernetes multus logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:13 +0000 UTC Normal Pod my-jaeger-collector-648fbf4b47-nndkr.spec.containers{jaeger-collector} Pulling Pulling image "test" kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:13 +0000 UTC Warning Pod my-jaeger-collector-648fbf4b47-nndkr.spec.containers{jaeger-collector} Failed Failed to pull image "test": reading manifest latest in docker.io/library/test: requested access to the resource is denied kubelet logger.go:42: 07:45:13 | set-custom-img | 2024-06-10 07:45:13 +0000 UTC Warning Pod my-jaeger-collector-648fbf4b47-nndkr.spec.containers{jaeger-collector} Failed Error: ErrImagePull kubelet logger.go:42: 07:45:14 | set-custom-img | Deleting namespace: kuttl-test-sharp-airedale === CONT kuttl/harness/collector-otlp-production-http logger.go:42: 07:45:20 | collector-otlp-production-http | Creating namespace: kuttl-test-driven-hermit logger.go:42: 07:45:20 | collector-otlp-production-http/1-install | starting test step 1-install logger.go:42: 07:45:20 | collector-otlp-production-http/1-install | Jaeger:kuttl-test-driven-hermit/my-jaeger created logger.go:42: 07:45:56 | collector-otlp-production-http/1-install | test step completed 1-install logger.go:42: 07:45:56 | collector-otlp-production-http/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:45:56 | collector-otlp-production-http/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:45:58 | collector-otlp-production-http/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:46:04 | collector-otlp-production-http/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:46:05 | collector-otlp-production-http/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:46:05 | collector-otlp-production-http/2-smoke-test | job.batch/report-span created logger.go:42: 07:46:05 | collector-otlp-production-http/2-smoke-test | job.batch/check-span created logger.go:42: 07:46:17 | collector-otlp-production-http/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:46:17 | collector-otlp-production-http | collector-otlp-production-http events from ns kuttl-test-driven-hermit: logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:26 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestdrivenhermitmyjaeger-1-5d46d9fbf8 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestdrivenhermitmyjaeger-1-5d46d9fbnh98q replicaset-controller logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdrivenhermitmyjaeger-1-5d46d9fbnh98q Binding Scheduled Successfully assigned kuttl-test-driven-hermit/elasticsearch-cdm-kuttltestdrivenhermitmyjaeger-1-5d46d9fbnh98q to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdrivenhermitmyjaeger-1-5d46d9fbnh98q AddedInterface Add eth0 [10.131.0.66/23] from ovn-kubernetes multus logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdrivenhermitmyjaeger-1-5d46d9fbnh98q.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdrivenhermitmyjaeger-1-5d46d9fbnh98q.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:26 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestdrivenhermitmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestdrivenhermitmyjaeger-1-5d46d9fbf8 to 1 deployment-controller logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdrivenhermitmyjaeger-1-5d46d9fbnh98q.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdrivenhermitmyjaeger-1-5d46d9fbnh98q.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdrivenhermitmyjaeger-1-5d46d9fbnh98q.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdrivenhermitmyjaeger-1-5d46d9fbnh98q.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:36 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestdrivenhermitmyjaeger-1-5d46d9fbnh98q.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:41 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestdrivenhermitmyjaeger-1-5d46d9fbnh98q.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:53 +0000 UTC Normal Pod my-jaeger-collector-6496d8c995-6h4w9 Binding Scheduled Successfully assigned kuttl-test-driven-hermit/my-jaeger-collector-6496d8c995-6h4w9 to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:53 +0000 UTC Normal Pod my-jaeger-collector-6496d8c995-6h4w9 AddedInterface Add eth0 [10.129.2.77/23] from ovn-kubernetes multus logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:53 +0000 UTC Normal Pod my-jaeger-collector-6496d8c995-6h4w9.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" already present on machine kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:53 +0000 UTC Normal Pod my-jaeger-collector-6496d8c995-6h4w9.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:53 +0000 UTC Normal Pod my-jaeger-collector-6496d8c995-6h4w9.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:53 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-6496d8c995 SuccessfulCreate Created pod: my-jaeger-collector-6496d8c995-6h4w9 replicaset-controller logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:53 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-6496d8c995 to 1 deployment-controller logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:53 +0000 UTC Normal Pod my-jaeger-query-77c99d59-z92gb Binding Scheduled Successfully assigned kuttl-test-driven-hermit/my-jaeger-query-77c99d59-z92gb to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:53 +0000 UTC Normal Pod my-jaeger-query-77c99d59-z92gb AddedInterface Add eth0 [10.128.2.66/23] from ovn-kubernetes multus logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:53 +0000 UTC Normal Pod my-jaeger-query-77c99d59-z92gb.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:53 +0000 UTC Normal Pod my-jaeger-query-77c99d59-z92gb.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:53 +0000 UTC Normal Pod my-jaeger-query-77c99d59-z92gb.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:53 +0000 UTC Normal Pod my-jaeger-query-77c99d59-z92gb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:53 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-77c99d59 SuccessfulCreate Created pod: my-jaeger-query-77c99d59-z92gb replicaset-controller logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:53 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-77c99d59 to 1 deployment-controller logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:54 +0000 UTC Normal Pod my-jaeger-query-77c99d59-z92gb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:54 +0000 UTC Normal Pod my-jaeger-query-77c99d59-z92gb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:54 +0000 UTC Normal Pod my-jaeger-query-77c99d59-z92gb.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:54 +0000 UTC Normal Pod my-jaeger-query-77c99d59-z92gb.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:54 +0000 UTC Normal Pod my-jaeger-query-77c99d59-z92gb.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:59 +0000 UTC Normal Pod my-jaeger-query-77c99d59-z92gb.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:59 +0000 UTC Normal Pod my-jaeger-query-77c99d59-z92gb.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:59 +0000 UTC Normal Pod my-jaeger-query-77c99d59-z92gb.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:59 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-77c99d59 SuccessfulDelete Deleted pod: my-jaeger-query-77c99d59-z92gb replicaset-controller logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:45:59 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-77c99d59 to 0 from 1 deployment-controller logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:00 +0000 UTC Normal Pod my-jaeger-query-7fbcb5787c-q7s7s Binding Scheduled Successfully assigned kuttl-test-driven-hermit/my-jaeger-query-7fbcb5787c-q7s7s to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:00 +0000 UTC Normal Pod my-jaeger-query-7fbcb5787c-q7s7s AddedInterface Add eth0 [10.128.2.67/23] from ovn-kubernetes multus logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:00 +0000 UTC Normal Pod my-jaeger-query-7fbcb5787c-q7s7s.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:00 +0000 UTC Normal Pod my-jaeger-query-7fbcb5787c-q7s7s.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:00 +0000 UTC Normal Pod my-jaeger-query-7fbcb5787c-q7s7s.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:00 +0000 UTC Normal Pod my-jaeger-query-7fbcb5787c-q7s7s.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:00 +0000 UTC Normal Pod my-jaeger-query-7fbcb5787c-q7s7s.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:00 +0000 UTC Normal Pod my-jaeger-query-7fbcb5787c-q7s7s.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:00 +0000 UTC Normal Pod my-jaeger-query-7fbcb5787c-q7s7s.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:00 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7fbcb5787c SuccessfulCreate Created pod: my-jaeger-query-7fbcb5787c-q7s7s replicaset-controller logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:00 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7fbcb5787c to 1 deployment-controller logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:01 +0000 UTC Normal Pod my-jaeger-query-7fbcb5787c-q7s7s.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:01 +0000 UTC Normal Pod my-jaeger-query-7fbcb5787c-q7s7s.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:05 +0000 UTC Normal Pod check-span-z84mr Binding Scheduled Successfully assigned kuttl-test-driven-hermit/check-span-z84mr to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:05 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-z84mr job-controller logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:05 +0000 UTC Normal Pod report-span-nd2nh Binding Scheduled Successfully assigned kuttl-test-driven-hermit/report-span-nd2nh to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:05 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-nd2nh job-controller logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:06 +0000 UTC Normal Pod check-span-z84mr AddedInterface Add eth0 [10.129.2.79/23] from ovn-kubernetes multus logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:06 +0000 UTC Normal Pod check-span-z84mr.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:06 +0000 UTC Normal Pod check-span-z84mr.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 306ms (306ms including waiting) kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:06 +0000 UTC Normal Pod check-span-z84mr.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:06 +0000 UTC Normal Pod check-span-z84mr.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:06 +0000 UTC Normal Pod report-span-nd2nh AddedInterface Add eth0 [10.129.2.78/23] from ovn-kubernetes multus logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:06 +0000 UTC Normal Pod report-span-nd2nh.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:06 +0000 UTC Normal Pod report-span-nd2nh.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 315ms (315ms including waiting) kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:06 +0000 UTC Normal Pod report-span-nd2nh.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:06 +0000 UTC Normal Pod report-span-nd2nh.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:08 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:08 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:08 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:46:17 | collector-otlp-production-http | 2024-06-10 07:46:17 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:46:17 | collector-otlp-production-http | Deleting namespace: kuttl-test-driven-hermit === CONT kuttl/harness/collector-otlp-production-grpc logger.go:42: 07:46:30 | collector-otlp-production-grpc | Creating namespace: kuttl-test-alive-bull logger.go:42: 07:46:30 | collector-otlp-production-grpc/1-install | starting test step 1-install logger.go:42: 07:46:30 | collector-otlp-production-grpc/1-install | Jaeger:kuttl-test-alive-bull/my-jaeger created logger.go:42: 07:47:06 | collector-otlp-production-grpc/1-install | test step completed 1-install logger.go:42: 07:47:06 | collector-otlp-production-grpc/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:47:06 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:47:08 | collector-otlp-production-grpc/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:47:14 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:47:15 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:47:15 | collector-otlp-production-grpc/2-smoke-test | job.batch/report-span created logger.go:42: 07:47:15 | collector-otlp-production-grpc/2-smoke-test | job.batch/check-span created logger.go:42: 07:47:36 | collector-otlp-production-grpc/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:47:36 | collector-otlp-production-grpc | collector-otlp-production-grpc events from ns kuttl-test-alive-bull: logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:46:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestalivebullmyjaeger-1-68cfb8699d-2m6sp Binding Scheduled Successfully assigned kuttl-test-alive-bull/elasticsearch-cdm-kuttltestalivebullmyjaeger-1-68cfb8699d-2m6sp to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:46:36 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestalivebullmyjaeger-1-68cfb8699d SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestalivebullmyjaeger-1-68cfb8699d-2m6sp replicaset-controller logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:46:36 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestalivebullmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestalivebullmyjaeger-1-68cfb8699d to 1 deployment-controller logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:46:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestalivebullmyjaeger-1-68cfb8699d-2m6sp AddedInterface Add eth0 [10.131.0.67/23] from ovn-kubernetes multus logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:46:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestalivebullmyjaeger-1-68cfb8699d-2m6sp.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:46:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestalivebullmyjaeger-1-68cfb8699d-2m6sp.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:46:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestalivebullmyjaeger-1-68cfb8699d-2m6sp.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:46:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestalivebullmyjaeger-1-68cfb8699d-2m6sp.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:46:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestalivebullmyjaeger-1-68cfb8699d-2m6sp.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:46:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestalivebullmyjaeger-1-68cfb8699d-2m6sp.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:46:47 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestalivebullmyjaeger-1-68cfb8699d-2m6sp.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:46:52 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestalivebullmyjaeger-1-68cfb8699d-2m6sp.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:03 +0000 UTC Normal Pod my-jaeger-collector-54656fd6d6-bpt9f Binding Scheduled Successfully assigned kuttl-test-alive-bull/my-jaeger-collector-54656fd6d6-bpt9f to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:03 +0000 UTC Normal Pod my-jaeger-collector-54656fd6d6-bpt9f AddedInterface Add eth0 [10.129.2.80/23] from ovn-kubernetes multus logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:03 +0000 UTC Normal Pod my-jaeger-collector-54656fd6d6-bpt9f.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" already present on machine kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:03 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-54656fd6d6 SuccessfulCreate Created pod: my-jaeger-collector-54656fd6d6-bpt9f replicaset-controller logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:03 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-54656fd6d6 to 1 deployment-controller logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:03 +0000 UTC Normal Pod my-jaeger-query-99985675f-7rlrw Binding Scheduled Successfully assigned kuttl-test-alive-bull/my-jaeger-query-99985675f-7rlrw to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:03 +0000 UTC Normal Pod my-jaeger-query-99985675f-7rlrw AddedInterface Add eth0 [10.128.2.68/23] from ovn-kubernetes multus logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:03 +0000 UTC Normal Pod my-jaeger-query-99985675f-7rlrw.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:03 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-99985675f SuccessfulCreate Created pod: my-jaeger-query-99985675f-7rlrw replicaset-controller logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:03 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-99985675f to 1 deployment-controller logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:04 +0000 UTC Normal Pod my-jaeger-collector-54656fd6d6-bpt9f.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:04 +0000 UTC Normal Pod my-jaeger-collector-54656fd6d6-bpt9f.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:04 +0000 UTC Normal Pod my-jaeger-query-99985675f-7rlrw.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:04 +0000 UTC Normal Pod my-jaeger-query-99985675f-7rlrw.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:04 +0000 UTC Normal Pod my-jaeger-query-99985675f-7rlrw.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:04 +0000 UTC Normal Pod my-jaeger-query-99985675f-7rlrw.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:04 +0000 UTC Normal Pod my-jaeger-query-99985675f-7rlrw.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:04 +0000 UTC Normal Pod my-jaeger-query-99985675f-7rlrw.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:04 +0000 UTC Normal Pod my-jaeger-query-99985675f-7rlrw.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:04 +0000 UTC Normal Pod my-jaeger-query-99985675f-7rlrw.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:09 +0000 UTC Normal Pod my-jaeger-query-99985675f-7rlrw.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:09 +0000 UTC Normal Pod my-jaeger-query-99985675f-7rlrw.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:09 +0000 UTC Normal Pod my-jaeger-query-99985675f-7rlrw.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:09 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-99985675f SuccessfulDelete Deleted pod: my-jaeger-query-99985675f-7rlrw replicaset-controller logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:09 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-99985675f to 0 from 1 deployment-controller logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:10 +0000 UTC Normal Pod my-jaeger-query-84b676cb9d-2r8rl Binding Scheduled Successfully assigned kuttl-test-alive-bull/my-jaeger-query-84b676cb9d-2r8rl to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:10 +0000 UTC Normal Pod my-jaeger-query-84b676cb9d-2r8rl AddedInterface Add eth0 [10.128.2.69/23] from ovn-kubernetes multus logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:10 +0000 UTC Normal Pod my-jaeger-query-84b676cb9d-2r8rl.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:10 +0000 UTC Normal Pod my-jaeger-query-84b676cb9d-2r8rl.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:10 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-84b676cb9d SuccessfulCreate Created pod: my-jaeger-query-84b676cb9d-2r8rl replicaset-controller logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:10 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-84b676cb9d to 1 deployment-controller logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:11 +0000 UTC Normal Pod my-jaeger-query-84b676cb9d-2r8rl.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:11 +0000 UTC Normal Pod my-jaeger-query-84b676cb9d-2r8rl.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:11 +0000 UTC Normal Pod my-jaeger-query-84b676cb9d-2r8rl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:11 +0000 UTC Normal Pod my-jaeger-query-84b676cb9d-2r8rl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:11 +0000 UTC Normal Pod my-jaeger-query-84b676cb9d-2r8rl.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:11 +0000 UTC Normal Pod my-jaeger-query-84b676cb9d-2r8rl.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:11 +0000 UTC Normal Pod my-jaeger-query-84b676cb9d-2r8rl.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:15 +0000 UTC Normal Pod check-span-ttdsw Binding Scheduled Successfully assigned kuttl-test-alive-bull/check-span-ttdsw to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:15 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-ttdsw job-controller logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:15 +0000 UTC Normal Pod report-span-pt229 Binding Scheduled Successfully assigned kuttl-test-alive-bull/report-span-pt229 to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:15 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-pt229 job-controller logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:16 +0000 UTC Normal Pod check-span-ttdsw AddedInterface Add eth0 [10.129.2.82/23] from ovn-kubernetes multus logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:16 +0000 UTC Normal Pod check-span-ttdsw.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:16 +0000 UTC Normal Pod check-span-ttdsw.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 360ms (360ms including waiting) kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:16 +0000 UTC Normal Pod check-span-ttdsw.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:16 +0000 UTC Normal Pod check-span-ttdsw.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:16 +0000 UTC Normal Pod report-span-pt229 AddedInterface Add eth0 [10.129.2.81/23] from ovn-kubernetes multus logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:16 +0000 UTC Normal Pod report-span-pt229.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:16 +0000 UTC Normal Pod report-span-pt229.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 343ms (343ms including waiting) kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:17 +0000 UTC Normal Pod report-span-pt229.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:17 +0000 UTC Normal Pod report-span-pt229.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:18 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:18 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:18 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:33 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:33 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-54656fd6d6-bpt9f horizontal-pod-autoscaler logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:33 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:47:36 | collector-otlp-production-grpc | 2024-06-10 07:47:35 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:47:36 | collector-otlp-production-grpc | Deleting namespace: kuttl-test-alive-bull === CONT kuttl/harness/collector-autoscale logger.go:42: 07:47:48 | collector-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:47:48 | collector-autoscale | Creating namespace: kuttl-test-primary-lemur logger.go:42: 07:47:48 | collector-autoscale/1-install | starting test step 1-install logger.go:42: 07:47:48 | collector-autoscale/1-install | Jaeger:kuttl-test-primary-lemur/simple-prod created logger.go:42: 07:48:26 | collector-autoscale/1-install | test step completed 1-install logger.go:42: 07:48:26 | collector-autoscale/2- | starting test step 2- logger.go:42: 07:48:26 | collector-autoscale/2- | test step completed 2- logger.go:42: 07:48:26 | collector-autoscale | collector-autoscale events from ns kuttl-test-primary-lemur: logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:47:55 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestprimarylemursimpleprod-1-98dcd847b SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestprimarylemursimpleprod-1-98dcd88vwgk replicaset-controller logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:47:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprimarylemursimpleprod-1-98dcd88vwgk Binding Scheduled Successfully assigned kuttl-test-primary-lemur/elasticsearch-cdm-kuttltestprimarylemursimpleprod-1-98dcd88vwgk to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:47:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprimarylemursimpleprod-1-98dcd88vwgk AddedInterface Add eth0 [10.129.2.83/23] from ovn-kubernetes multus logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:47:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprimarylemursimpleprod-1-98dcd88vwgk.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:47:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprimarylemursimpleprod-1-98dcd88vwgk.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:47:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprimarylemursimpleprod-1-98dcd88vwgk.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:47:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprimarylemursimpleprod-1-98dcd88vwgk.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:47:55 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestprimarylemursimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestprimarylemursimpleprod-1-98dcd847b to 1 deployment-controller logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:47:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprimarylemursimpleprod-1-98dcd88vwgk.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:47:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprimarylemursimpleprod-1-98dcd88vwgk.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:05 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestprimarylemursimpleprod-1-98dcd88vwgk.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:10 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestprimarylemursimpleprod-1-98dcd88vwgk.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:22 +0000 UTC Normal Pod simple-prod-collector-5dbc48d689-2bcwp Binding Scheduled Successfully assigned kuttl-test-primary-lemur/simple-prod-collector-5dbc48d689-2bcwp to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:22 +0000 UTC Normal Pod simple-prod-collector-5dbc48d689-2bcwp AddedInterface Add eth0 [10.131.0.68/23] from ovn-kubernetes multus logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:22 +0000 UTC Normal Pod simple-prod-collector-5dbc48d689-2bcwp.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:c545b5167d63ae3e3ff7f16eed02b51172332e7b28ea4634ad7045f7cbe95f45" already present on machine kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:22 +0000 UTC Normal Pod simple-prod-collector-5dbc48d689-2bcwp.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:22 +0000 UTC Normal Pod simple-prod-collector-5dbc48d689-2bcwp.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:22 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5dbc48d689 SuccessfulCreate Created pod: simple-prod-collector-5dbc48d689-2bcwp replicaset-controller logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:22 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-5dbc48d689 to 1 deployment-controller logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:22 +0000 UTC Normal Pod simple-prod-query-6f76fd48db-vm7jq Binding Scheduled Successfully assigned kuttl-test-primary-lemur/simple-prod-query-6f76fd48db-vm7jq to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:22 +0000 UTC Warning Pod simple-prod-query-6f76fd48db-vm7jq FailedMount MountVolume.SetUp failed for volume "simple-prod-ui-oauth-proxy-tls" : secret "simple-prod-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:22 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6f76fd48db SuccessfulCreate Created pod: simple-prod-query-6f76fd48db-vm7jq replicaset-controller logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:22 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6f76fd48db to 1 deployment-controller logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:23 +0000 UTC Normal Pod simple-prod-query-6f76fd48db-vm7jq AddedInterface Add eth0 [10.128.2.70/23] from ovn-kubernetes multus logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:23 +0000 UTC Normal Pod simple-prod-query-6f76fd48db-vm7jq.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:95615560040a62e07fac86b347b68d9b5b449f293184e87c260e4edc00f918b4" already present on machine kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:23 +0000 UTC Normal Pod simple-prod-query-6f76fd48db-vm7jq.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:23 +0000 UTC Normal Pod simple-prod-query-6f76fd48db-vm7jq.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:23 +0000 UTC Normal Pod simple-prod-query-6f76fd48db-vm7jq.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:23 +0000 UTC Normal Pod simple-prod-query-6f76fd48db-vm7jq.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:23 +0000 UTC Normal Pod simple-prod-query-6f76fd48db-vm7jq.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:23 +0000 UTC Normal Pod simple-prod-query-6f76fd48db-vm7jq.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5783e9fd32d80d78ed8b6842eb6362b730b13f2ff098cf305e8887dd22e13477" already present on machine kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:23 +0000 UTC Normal Pod simple-prod-query-6f76fd48db-vm7jq.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:48:26 | collector-autoscale | 2024-06-10 07:48:23 +0000 UTC Normal Pod simple-prod-query-6f76fd48db-vm7jq.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:48:26 | collector-autoscale | Deleting namespace: kuttl-test-primary-lemur === CONT kuttl/harness/collector-otlp-allinone-grpc logger.go:42: 07:48:32 | collector-otlp-allinone-grpc | Creating namespace: kuttl-test-national-goldfish logger.go:42: 07:48:32 | collector-otlp-allinone-grpc/0-install | starting test step 0-install logger.go:42: 07:48:32 | collector-otlp-allinone-grpc/0-install | Jaeger:kuttl-test-national-goldfish/my-jaeger created logger.go:42: 07:48:37 | collector-otlp-allinone-grpc/0-install | test step completed 0-install logger.go:42: 07:48:37 | collector-otlp-allinone-grpc/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:48:37 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:48:39 | collector-otlp-allinone-grpc/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:48:46 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:48:46 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:48:47 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/report-span created logger.go:42: 07:48:47 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/check-span created logger.go:42: 07:49:07 | collector-otlp-allinone-grpc/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | collector-otlp-allinone-grpc events from ns kuttl-test-national-goldfish: logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:35 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-78f78959fc to 1 deployment-controller logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:36 +0000 UTC Normal Pod my-jaeger-78f78959fc-89b72 Binding Scheduled Successfully assigned kuttl-test-national-goldfish/my-jaeger-78f78959fc-89b72 to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:36 +0000 UTC Normal Pod my-jaeger-78f78959fc-89b72 AddedInterface Add eth0 [10.128.2.71/23] from ovn-kubernetes multus logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:36 +0000 UTC Normal Pod my-jaeger-78f78959fc-89b72.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:36 +0000 UTC Normal Pod my-jaeger-78f78959fc-89b72.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:36 +0000 UTC Normal Pod my-jaeger-78f78959fc-89b72.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:36 +0000 UTC Normal Pod my-jaeger-78f78959fc-89b72.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:36 +0000 UTC Normal Pod my-jaeger-78f78959fc-89b72.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:36 +0000 UTC Normal Pod my-jaeger-78f78959fc-89b72.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:36 +0000 UTC Normal ReplicaSet.apps my-jaeger-78f78959fc SuccessfulCreate Created pod: my-jaeger-78f78959fc-89b72 replicaset-controller logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:43 +0000 UTC Normal Pod my-jaeger-78f78959fc-89b72.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:43 +0000 UTC Normal Pod my-jaeger-78f78959fc-89b72.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:43 +0000 UTC Normal ReplicaSet.apps my-jaeger-78f78959fc SuccessfulDelete Deleted pod: my-jaeger-78f78959fc-89b72 replicaset-controller logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:43 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-78f78959fc to 0 from 1 deployment-controller logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:44 +0000 UTC Normal Pod my-jaeger-596c4948c7-2q792 Binding Scheduled Successfully assigned kuttl-test-national-goldfish/my-jaeger-596c4948c7-2q792 to ip-10-0-121-113.us-east-2.compute.internal default-scheduler logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:44 +0000 UTC Normal ReplicaSet.apps my-jaeger-596c4948c7 SuccessfulCreate Created pod: my-jaeger-596c4948c7-2q792 replicaset-controller logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:44 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-596c4948c7 to 1 deployment-controller logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:45 +0000 UTC Normal Pod my-jaeger-596c4948c7-2q792 AddedInterface Add eth0 [10.128.2.72/23] from ovn-kubernetes multus logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:45 +0000 UTC Normal Pod my-jaeger-596c4948c7-2q792.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:c166f01cff83b0c65c571abaf502d26defcf35c75ff9d0d9d95f895340b64db6" already present on machine kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:45 +0000 UTC Normal Pod my-jaeger-596c4948c7-2q792.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:45 +0000 UTC Normal Pod my-jaeger-596c4948c7-2q792.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:45 +0000 UTC Normal Pod my-jaeger-596c4948c7-2q792.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:9dd175ddce0d481b120665bf15c5946fdc4111617e37401995e8292e865f6ebf" already present on machine kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:45 +0000 UTC Normal Pod my-jaeger-596c4948c7-2q792.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:45 +0000 UTC Normal Pod my-jaeger-596c4948c7-2q792.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:47 +0000 UTC Normal Pod check-span-8hmzm Binding Scheduled Successfully assigned kuttl-test-national-goldfish/check-span-8hmzm to ip-10-0-70-218.us-east-2.compute.internal default-scheduler logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:47 +0000 UTC Normal Pod check-span-8hmzm AddedInterface Add eth0 [10.129.2.84/23] from ovn-kubernetes multus logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:47 +0000 UTC Normal Pod check-span-8hmzm.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:47 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-8hmzm job-controller logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:47 +0000 UTC Normal Pod report-span-q8lhp Binding Scheduled Successfully assigned kuttl-test-national-goldfish/report-span-q8lhp to ip-10-0-8-114.us-east-2.compute.internal default-scheduler logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:47 +0000 UTC Normal Pod report-span-q8lhp AddedInterface Add eth0 [10.131.0.69/23] from ovn-kubernetes multus logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:47 +0000 UTC Normal Pod report-span-q8lhp.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:47 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-q8lhp job-controller logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:48 +0000 UTC Normal Pod check-span-8hmzm.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 327ms (327ms including waiting) kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:48 +0000 UTC Normal Pod check-span-8hmzm.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:48 +0000 UTC Normal Pod check-span-8hmzm.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:48 +0000 UTC Normal Pod report-span-q8lhp.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 474ms (474ms including waiting) kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:48 +0000 UTC Normal Pod report-span-q8lhp.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:48:48 +0000 UTC Normal Pod report-span-q8lhp.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | 2024-06-10 07:49:07 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:49:07 | collector-otlp-allinone-grpc | Deleting namespace: kuttl-test-national-goldfish === CONT kuttl/harness/cassandra-spark logger.go:42: 07:49:19 | cassandra-spark | Ignoring 01-assert.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:49:19 | cassandra-spark | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:49:19 | cassandra-spark | Creating namespace: kuttl-test-holy-fox logger.go:42: 07:49:19 | cassandra-spark | cassandra-spark events from ns kuttl-test-holy-fox: logger.go:42: 07:49:19 | cassandra-spark | Deleting namespace: kuttl-test-holy-fox === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (339.01s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.82s) --- PASS: kuttl/harness/collector-otlp-allinone-http (39.34s) --- PASS: kuttl/harness/set-custom-img (48.26s) --- PASS: kuttl/harness/collector-otlp-production-http (69.80s) --- PASS: kuttl/harness/collector-otlp-production-grpc (78.38s) --- PASS: kuttl/harness/collector-autoscale (43.86s) --- PASS: kuttl/harness/collector-otlp-allinone-grpc (47.53s) --- PASS: kuttl/harness/cassandra-spark (5.84s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name miscellaneous --report --output /logs/artifacts/miscellaneous.xml ./artifacts/kuttl-report.xml time="2024-06-10T07:49:27Z" level=debug msg="Setting a new name for the test suites" time="2024-06-10T07:49:27Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-06-10T07:49:27Z" level=debug msg="normalizing test case names" time="2024-06-10T07:49:27Z" level=debug msg="miscellaneous/artifacts -> miscellaneous_artifacts" time="2024-06-10T07:49:27Z" level=debug msg="miscellaneous/collector-otlp-allinone-http -> miscellaneous_collector_otlp_allinone_http" time="2024-06-10T07:49:27Z" level=debug msg="miscellaneous/set-custom-img -> miscellaneous_set_custom_img" time="2024-06-10T07:49:27Z" level=debug msg="miscellaneous/collector-otlp-production-http -> miscellaneous_collector_otlp_production_http" time="2024-06-10T07:49:27Z" level=debug msg="miscellaneous/collector-otlp-production-grpc -> miscellaneous_collector_otlp_production_grpc" time="2024-06-10T07:49:27Z" level=debug msg="miscellaneous/collector-autoscale -> miscellaneous_collector_autoscale" time="2024-06-10T07:49:27Z" level=debug msg="miscellaneous/collector-otlp-allinone-grpc -> miscellaneous_collector_otlp_allinone_grpc" time="2024-06-10T07:49:27Z" level=debug msg="miscellaneous/cassandra-spark -> miscellaneous_cassandra_spark" +----------------------------------------------+--------+ | NAME | RESULT | +----------------------------------------------+--------+ | miscellaneous_artifacts | passed | | miscellaneous_collector_otlp_allinone_http | passed | | miscellaneous_set_custom_img | passed | | miscellaneous_collector_otlp_production_http | passed | | miscellaneous_collector_otlp_production_grpc | passed | | miscellaneous_collector_autoscale | passed | | miscellaneous_collector_otlp_allinone_grpc | passed | | miscellaneous_cassandra_spark | passed | +----------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests'