Cloning into '/tmp/jaeger-tests'... Switched to a new branch 'integration-tests' branch 'integration-tests' set up to track 'origin/integration-tests'. Installing kuttl Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/kubectl-kuttl https://github.com/kudobuilder/kuttl/releases/download/v0.15.0/kubectl-kuttl_0.15.0_linux_x86_64 KUBECONFIG file is: /tmp/kubeconfig-998185858 for suite in elasticsearch examples generate miscellaneous sidecar streaming ui upgrade; do \ make run-e2e-tests-$suite ; \ done make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh elasticsearch false true + '[' 3 -ne 3 ']' + test_suite_name=elasticsearch + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/elasticsearch.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-elasticsearch make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true \ KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ ./tests/e2e/elasticsearch/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-14-222249 True False 10m Cluster version is 4.18.0-0.nightly-2025-02-14-222249' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-14-222249 True False 10m Cluster version is 4.18.0-0.nightly-2025-02-14-222249' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/elasticsearch/render.sh ++ export SUITE_DIR=./tests/e2e/elasticsearch ++ SUITE_DIR=./tests/e2e/elasticsearch ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/gomplate https://github.com/hairyhenderson/gomplate/releases/download/v3.10.0/gomplate_linux-amd64-slim ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/yq https://github.com/mikefarah/yq/releases/download/v4.20.2/yq_linux_amd64 ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize Try 0... curl -sLo /tmp/kustomize.tar.gz https://github.com/kubernetes-sigs/kustomize/releases/download/kustomize%2Fv4.5.7/kustomize_v4.5.7_linux_amd64.tar.gz Try 1... curl -sLo /tmp/kustomize.tar.gz https://github.com/kubernetes-sigs/kustomize/releases/download/kustomize%2Fv4.5.7/kustomize_v4.5.7_linux_amd64.tar.gz Try 2... curl -sLo /tmp/kustomize.tar.gz https://github.com/kubernetes-sigs/kustomize/releases/download/kustomize%2Fv4.5.7/kustomize_v4.5.7_linux_amd64.tar.gz Try 3... curl -sLo /tmp/kustomize.tar.gz https://github.com/kubernetes-sigs/kustomize/releases/download/kustomize%2Fv4.5.7/kustomize_v4.5.7_linux_amd64.tar.gz Try 4... curl -sLo /tmp/kustomize.tar.gz https://github.com/kubernetes-sigs/kustomize/releases/download/kustomize%2Fv4.5.7/kustomize_v4.5.7_linux_amd64.tar.gz ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/elasticsearch ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + start_test es-from-aio-to-production + '[' 1 -ne 1 ']' + test_name=es-from-aio-to-production + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-from-aio-to-production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-from-aio-to-production\e[0m' Rendering files for test es-from-aio-to-production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-from-aio-to-production + cd es-from-aio-to-production + jaeger_name=my-jaeger + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 03 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=03 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./03-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch.redundancyPolicy="ZeroRedundancy"' ./03-install.yaml + render_smoke_test my-jaeger true 04 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=04 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./04-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test es-increasing-replicas + '[' 1 -ne 1 ']' + test_name=es-increasing-replicas + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-increasing-replicas' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-increasing-replicas\e[0m' Rendering files for test es-increasing-replicas + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-from-aio-to-production + '[' es-from-aio-to-production '!=' _build ']' + cd .. + mkdir -p es-increasing-replicas + cd es-increasing-replicas + jaeger_name=simple-prod + '[' true = true ']' + jaeger_deployment_mode=production_autoprovisioned + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.replicas=2 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.query.replicas=2 ./02-install.yaml + cp ./01-assert.yaml ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.replicas=2 ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .status.readyReplicas=2 ./02-assert.yaml + render_smoke_test simple-prod true 03 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=03 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./03-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + cp ./02-install.yaml ./04-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.elasticsearch.nodeCount=2 ./04-install.yaml + /tmp/jaeger-tests/bin/gomplate -f ./openshift-check-es-nodes.yaml.template -o ./05-check-es-nodes.yaml + '[' true = true ']' + skip_test es-index-cleaner-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-increasing-replicas + '[' es-increasing-replicas '!=' _build ']' + cd .. + rm -rf es-index-cleaner-upstream + warning 'es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_index_cleaner -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-index-cleaner-autoprov + '[' 1 -ne 1 ']' + test_name=es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-index-cleaner-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-index-cleaner-autoprov\e[0m' Rendering files for test es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-index-cleaner-autoprov + cd es-index-cleaner-autoprov + jaeger_name=test-es-index-cleaner-with-prefix + cronjob_name=test-es-index-cleaner-with-prefix-es-index-cleaner + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + cp ../../es-index-cleaner-upstream/04-assert.yaml ../../es-index-cleaner-upstream/README.md . + render_install_jaeger test-es-index-cleaner-with-prefix production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options.es.index-prefix=""' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.enabled=false ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.numberOfDays=0 ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.esIndexCleaner.schedule="*/1 * * * *"' ./01-install.yaml + render_report_spans test-es-index-cleaner-with-prefix true 5 00 true 02 + '[' 6 -ne 6 ']' + jaeger=test-es-index-cleaner-with-prefix + is_secured=true + number_of_spans=5 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + export JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=5 + DAYS=5 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + sed 's~enabled: false~enabled: true~gi' ./01-install.yaml + CRONJOB_NAME=test-es-index-cleaner-with-prefix-es-index-cleaner + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./04-wait-es-index-cleaner.yaml + /tmp/jaeger-tests/bin/gomplate -f ./01-install.yaml -o ./05-install.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 00 06 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=00 + test_step=06 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=test-es-index-cleaner-with-prefix-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=test-es-index-cleaner-with-prefix-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./06-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./06-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.17"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.17"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.17"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.17 ++ version_ge 5.8.17 5.4 +++ echo 5.8.17 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.17 == 5.8.17 + '[' -n '' ']' + skip_test es-index-cleaner-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-index-cleaner-autoprov + '[' es-index-cleaner-autoprov '!=' _build ']' + cd .. + rm -rf es-index-cleaner-managed + warning 'es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + start_test es-multiinstance + '[' 1 -ne 1 ']' + test_name=es-multiinstance + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-multiinstance' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-multiinstance\e[0m' Rendering files for test es-multiinstance + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-multiinstance + cd es-multiinstance + jaeger_name=instance-1 + render_install_jaeger instance-1 production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=instance-1 + JAEGER_NAME=instance-1 + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f ./03-create-second-instance.yaml.template -o 03-create-second-instance.yaml + '[' true = true ']' + skip_test es-rollover-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-rollover-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-multiinstance + '[' es-multiinstance '!=' _build ']' + cd .. + rm -rf es-rollover-upstream + warning 'es-rollover-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_rollover -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-rollover-autoprov + '[' 1 -ne 1 ']' + test_name=es-rollover-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-rollover-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-rollover-autoprov\e[0m' Rendering files for test es-rollover-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-rollover-autoprov + cd es-rollover-autoprov + cp ../../es-rollover-upstream/05-assert.yaml ../../es-rollover-upstream/05-install.yaml ../../es-rollover-upstream/README.md . + jaeger_name=my-jaeger + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_report_spans my-jaeger true 2 00 true 02 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 00 03 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=00 + test_step=03 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./03-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./03-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 01 04 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=01 + test_step=04 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=01 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./04-check-indices.yaml + JOB_NUMBER=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./04-assert.yaml + render_report_spans my-jaeger true 2 02 true 06 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=02 + ensure_reported_spans=true + test_step=06 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=02 + JOB_NUMBER=02 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./06-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./06-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 02 07 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=02 + test_step=07 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=02 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./07-check-indices.yaml + JOB_NUMBER=02 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./07-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' 03 08 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + job_number=03 + test_step=08 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=03 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./08-check-indices.yaml + JOB_NUMBER=03 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./08-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' 04 09 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + job_number=04 + test_step=09 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=04 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./09-check-indices.yaml + JOB_NUMBER=04 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./09-assert.yaml + render_report_spans my-jaeger true 2 03 true 10 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=03 + ensure_reported_spans=true + test_step=10 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=03 + JOB_NUMBER=03 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./10-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./10-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + CRONJOB_NAME=my-jaeger-es-rollover + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./11-wait-rollover.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-000002'\'',' 05 11 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-000002'\'',' + job_number=05 + test_step=11 + escape_command ''\''--name'\'', '\''jaeger-span-000002'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-000002'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-000002'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-000002'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=05 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./11-check-indices.yaml + JOB_NUMBER=05 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./11-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' 06 12 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + job_number=06 + test_step=12 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=06 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./12-check-indices.yaml + JOB_NUMBER=06 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./12-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.17"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.17"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.17"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.17 ++ version_ge 5.8.17 5.4 +++ echo 5.8.17 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.17 == 5.8.17 + '[' -n '' ']' + skip_test es-rollover-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-rollover-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-rollover-autoprov + '[' es-rollover-autoprov '!=' _build ']' + cd .. + rm -rf es-rollover-managed + warning 'es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + skip_test es-spark-dependencies 'This test is not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=es-spark-dependencies + message='This test is not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + rm -rf es-spark-dependencies + warning 'es-spark-dependencies: This test is not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-spark-dependencies: This test is not supported in OpenShift\e[0m' WAR: es-spark-dependencies: This test is not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running elasticsearch E2E tests' Running elasticsearch E2E tests + cd tests/e2e/elasticsearch/_build + set +e + KUBECONFIG=/tmp/kubeconfig-998185858 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/es-from-aio-to-production === PAUSE kuttl/harness/es-from-aio-to-production === RUN kuttl/harness/es-increasing-replicas === PAUSE kuttl/harness/es-increasing-replicas === RUN kuttl/harness/es-index-cleaner-autoprov === PAUSE kuttl/harness/es-index-cleaner-autoprov === RUN kuttl/harness/es-multiinstance === PAUSE kuttl/harness/es-multiinstance === RUN kuttl/harness/es-rollover-autoprov === PAUSE kuttl/harness/es-rollover-autoprov === RUN kuttl/harness/es-simple-prod === PAUSE kuttl/harness/es-simple-prod === CONT kuttl/harness/artifacts logger.go:42: 20:14:56 | artifacts | Creating namespace: kuttl-test-correct-troll logger.go:42: 20:14:56 | artifacts | artifacts events from ns kuttl-test-correct-troll: logger.go:42: 20:14:56 | artifacts | Deleting namespace: kuttl-test-correct-troll === CONT kuttl/harness/es-multiinstance logger.go:42: 20:15:03 | es-multiinstance | Ignoring 03-create-second-instance.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:15:03 | es-multiinstance | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:15:03 | es-multiinstance | Creating namespace: kuttl-test-concise-llama logger.go:42: 20:15:03 | es-multiinstance/0-clear-namespace | starting test step 0-clear-namespace logger.go:42: 20:15:03 | es-multiinstance/0-clear-namespace | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --ignore-not-found=true] logger.go:42: 20:15:03 | es-multiinstance/0-clear-namespace | test step completed 0-clear-namespace logger.go:42: 20:15:03 | es-multiinstance/1-install | starting test step 1-install logger.go:42: 20:15:03 | es-multiinstance/1-install | Jaeger:kuttl-test-concise-llama/instance-1 created logger.go:42: 20:15:59 | es-multiinstance/1-install | test step completed 1-install logger.go:42: 20:15:59 | es-multiinstance/2-create-namespace | starting test step 2-create-namespace logger.go:42: 20:15:59 | es-multiinstance/2-create-namespace | running command: [sh -c kubectl create namespace jaeger-e2e-multiinstance-test] logger.go:42: 20:15:59 | es-multiinstance/2-create-namespace | namespace/jaeger-e2e-multiinstance-test created logger.go:42: 20:15:59 | es-multiinstance/2-create-namespace | test step completed 2-create-namespace logger.go:42: 20:15:59 | es-multiinstance/3-create-second-instance | starting test step 3-create-second-instance logger.go:42: 20:15:59 | es-multiinstance/3-create-second-instance | running command: [sh -c kubectl apply -f ./01-install.yaml -n jaeger-e2e-multiinstance-test] logger.go:42: 20:15:59 | es-multiinstance/3-create-second-instance | jaeger.jaegertracing.io/instance-1 created logger.go:42: 20:15:59 | es-multiinstance/3-create-second-instance | running command: [sh -c /tmp/jaeger-tests/bin/kubectl-kuttl assert ./01-assert.yaml -n jaeger-e2e-multiinstance-test --timeout 1000] logger.go:42: 20:16:48 | es-multiinstance/3-create-second-instance | assert is valid logger.go:42: 20:16:49 | es-multiinstance/3-create-second-instance | test step completed 3-create-second-instance logger.go:42: 20:16:49 | es-multiinstance/4-check-secrets | starting test step 4-check-secrets logger.go:42: 20:16:49 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n $NAMESPACE > secret1] logger.go:42: 20:16:49 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n jaeger-e2e-multiinstance-test > secret2] logger.go:42: 20:16:49 | es-multiinstance/4-check-secrets | running command: [sh -c cmp --silent secret1 secret2 || exit 0] logger.go:42: 20:16:49 | es-multiinstance/4-check-secrets | test step completed 4-check-secrets logger.go:42: 20:16:49 | es-multiinstance/5-delete | starting test step 5-delete logger.go:42: 20:16:49 | es-multiinstance/5-delete | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --wait=false] logger.go:42: 20:16:49 | es-multiinstance/5-delete | namespace "jaeger-e2e-multiinstance-test" deleted logger.go:42: 20:16:49 | es-multiinstance/5-delete | test step completed 5-delete logger.go:42: 20:16:49 | es-multiinstance | es-multiinstance events from ns kuttl-test-concise-llama: logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:10 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestconcisellamainstance1-1-54cc5ddccd SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestconcisellamainstance1-1-54cc5dds9hgd replicaset-controller logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestconcisellamainstance1-1-54cc5dds9hgd Binding Scheduled Successfully assigned kuttl-test-concise-llama/elasticsearch-cdm-kuttltestconcisellamainstance1-1-54cc5dds9hgd to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:10 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestconcisellamainstance1-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestconcisellamainstance1-1-54cc5ddccd to 1 deployment-controller logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestconcisellamainstance1-1-54cc5dds9hgd AddedInterface Add eth0 [10.129.2.25/23] from ovn-kubernetes multus logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestconcisellamainstance1-1-54cc5dds9hgd.spec.containers{elasticsearch} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:5117d0aee260f74d6ef77c7c71ed073b1ad45b456f58894539b5697bec878310" kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestconcisellamainstance1-1-54cc5dds9hgd.spec.containers{elasticsearch} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:5117d0aee260f74d6ef77c7c71ed073b1ad45b456f58894539b5697bec878310" in 9.286s (9.286s including waiting). Image size: 538303933 bytes. kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestconcisellamainstance1-1-54cc5dds9hgd.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestconcisellamainstance1-1-54cc5dds9hgd.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestconcisellamainstance1-1-54cc5dds9hgd.spec.containers{proxy} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d63d612cdca72f64c45960a9b30bcfd0c746daabc17407df92c62694d5bc4bb6" kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestconcisellamainstance1-1-54cc5dds9hgd.spec.containers{proxy} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d63d612cdca72f64c45960a9b30bcfd0c746daabc17407df92c62694d5bc4bb6" in 2.71s (2.71s including waiting). Image size: 286162050 bytes. kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestconcisellamainstance1-1-54cc5dds9hgd.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestconcisellamainstance1-1-54cc5dds9hgd.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:31 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestconcisellamainstance1-1-54cc5dds9hgd.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:36 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestconcisellamainstance1-1-54cc5dds9hgd.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:46 +0000 UTC Normal Pod instance-1-collector-544c4f8795-55wmb Binding Scheduled Successfully assigned kuttl-test-concise-llama/instance-1-collector-544c4f8795-55wmb to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:46 +0000 UTC Normal ReplicaSet.apps instance-1-collector-544c4f8795 SuccessfulCreate Created pod: instance-1-collector-544c4f8795-55wmb replicaset-controller logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:46 +0000 UTC Normal Deployment.apps instance-1-collector ScalingReplicaSet Scaled up replica set instance-1-collector-544c4f8795 to 1 deployment-controller logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:47 +0000 UTC Normal Pod instance-1-collector-544c4f8795-55wmb AddedInterface Add eth0 [10.131.0.17/23] from ovn-kubernetes multus logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:47 +0000 UTC Normal Pod instance-1-collector-544c4f8795-55wmb.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:47 +0000 UTC Normal Pod instance-1-query-864dc988b9-nb842 Binding Scheduled Successfully assigned kuttl-test-concise-llama/instance-1-query-864dc988b9-nb842 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:47 +0000 UTC Normal Pod instance-1-query-864dc988b9-nb842 AddedInterface Add eth0 [10.128.2.16/23] from ovn-kubernetes multus logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:47 +0000 UTC Normal Pod instance-1-query-864dc988b9-nb842.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:47 +0000 UTC Normal ReplicaSet.apps instance-1-query-864dc988b9 SuccessfulCreate Created pod: instance-1-query-864dc988b9-nb842 replicaset-controller logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:47 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-864dc988b9 to 1 deployment-controller logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:50 +0000 UTC Normal Pod instance-1-query-864dc988b9-nb842.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" in 2.816s (2.816s including waiting). Image size: 142020742 bytes. kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:50 +0000 UTC Normal Pod instance-1-query-864dc988b9-nb842.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:50 +0000 UTC Normal Pod instance-1-query-864dc988b9-nb842.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:50 +0000 UTC Normal Pod instance-1-query-864dc988b9-nb842.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:51 +0000 UTC Normal Pod instance-1-collector-544c4f8795-55wmb.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" in 3.457s (3.457s including waiting). Image size: 139779827 bytes. kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:51 +0000 UTC Normal Pod instance-1-collector-544c4f8795-55wmb.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:51 +0000 UTC Normal Pod instance-1-collector-544c4f8795-55wmb.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:54 +0000 UTC Normal Pod instance-1-query-864dc988b9-nb842.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" in 4.26s (4.26s including waiting). Image size: 438322369 bytes. kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:54 +0000 UTC Normal Pod instance-1-query-864dc988b9-nb842.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:54 +0000 UTC Normal Pod instance-1-query-864dc988b9-nb842.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:54 +0000 UTC Normal Pod instance-1-query-864dc988b9-nb842.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:57 +0000 UTC Normal Pod instance-1-query-864dc988b9-nb842.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" in 2.356s (2.356s including waiting). Image size: 112614125 bytes. kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:57 +0000 UTC Normal Pod instance-1-query-864dc988b9-nb842.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:15:57 +0000 UTC Normal Pod instance-1-query-864dc988b9-nb842.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:16:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:16:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:16:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:16:29 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod instance-1-collector-544c4f8795-55wmb horizontal-pod-autoscaler logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:16:44 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod instance-1-collector-544c4f8795-55wmb horizontal-pod-autoscaler logger.go:42: 20:16:49 | es-multiinstance | 2025-02-17 20:16:44 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod instance-1-collector-544c4f8795-55wmb horizontal-pod-autoscaler logger.go:42: 20:16:49 | es-multiinstance | Deleting namespace: kuttl-test-concise-llama === CONT kuttl/harness/es-simple-prod logger.go:42: 20:16:57 | es-simple-prod | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:16:57 | es-simple-prod | Creating namespace: kuttl-test-star-spider logger.go:42: 20:16:57 | es-simple-prod | es-simple-prod events from ns kuttl-test-star-spider: logger.go:42: 20:16:57 | es-simple-prod | Deleting namespace: kuttl-test-star-spider === CONT kuttl/harness/es-rollover-autoprov logger.go:42: 20:17:03 | es-rollover-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:17:03 | es-rollover-autoprov | Creating namespace: kuttl-test-present-escargot logger.go:42: 20:17:03 | es-rollover-autoprov/1-install | starting test step 1-install logger.go:42: 20:17:04 | es-rollover-autoprov/1-install | Jaeger:kuttl-test-present-escargot/my-jaeger created logger.go:42: 20:17:41 | es-rollover-autoprov/1-install | test step completed 1-install logger.go:42: 20:17:41 | es-rollover-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 20:17:41 | es-rollover-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 20:17:42 | es-rollover-autoprov/2-report-spans | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:17:50 | es-rollover-autoprov/2-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 20:17:51 | es-rollover-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 20:17:51 | es-rollover-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 20:18:17 | es-rollover-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 20:18:17 | es-rollover-autoprov/3-check-indices | starting test step 3-check-indices logger.go:42: 20:18:18 | es-rollover-autoprov/3-check-indices | Job:kuttl-test-present-escargot/00-check-indices created logger.go:42: 20:18:23 | es-rollover-autoprov/3-check-indices | test step completed 3-check-indices logger.go:42: 20:18:23 | es-rollover-autoprov/4-check-indices | starting test step 4-check-indices logger.go:42: 20:18:23 | es-rollover-autoprov/4-check-indices | Job:kuttl-test-present-escargot/01-check-indices created logger.go:42: 20:18:27 | es-rollover-autoprov/4-check-indices | test step completed 4-check-indices logger.go:42: 20:18:27 | es-rollover-autoprov/5-install | starting test step 5-install logger.go:42: 20:18:27 | es-rollover-autoprov/5-install | Jaeger:kuttl-test-present-escargot/my-jaeger updated logger.go:42: 20:18:34 | es-rollover-autoprov/5-install | test step completed 5-install logger.go:42: 20:18:34 | es-rollover-autoprov/6-report-spans | starting test step 6-report-spans logger.go:42: 20:18:34 | es-rollover-autoprov/6-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 20:18:44 | es-rollover-autoprov/6-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=02 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-02-job.yaml] logger.go:42: 20:18:44 | es-rollover-autoprov/6-report-spans | running command: [sh -c kubectl apply -f report-span-02-job.yaml -n $NAMESPACE] logger.go:42: 20:18:44 | es-rollover-autoprov/6-report-spans | job.batch/02-report-span created logger.go:42: 20:19:11 | es-rollover-autoprov/6-report-spans | test step completed 6-report-spans logger.go:42: 20:19:11 | es-rollover-autoprov/7-check-indices | starting test step 7-check-indices logger.go:42: 20:19:11 | es-rollover-autoprov/7-check-indices | Job:kuttl-test-present-escargot/02-check-indices created logger.go:42: 20:19:16 | es-rollover-autoprov/7-check-indices | test step completed 7-check-indices logger.go:42: 20:19:16 | es-rollover-autoprov/8-check-indices | starting test step 8-check-indices logger.go:42: 20:19:16 | es-rollover-autoprov/8-check-indices | Job:kuttl-test-present-escargot/03-check-indices created logger.go:42: 20:19:20 | es-rollover-autoprov/8-check-indices | test step completed 8-check-indices logger.go:42: 20:19:20 | es-rollover-autoprov/9-check-indices | starting test step 9-check-indices logger.go:42: 20:19:20 | es-rollover-autoprov/9-check-indices | Job:kuttl-test-present-escargot/04-check-indices created logger.go:42: 20:19:24 | es-rollover-autoprov/9-check-indices | test step completed 9-check-indices logger.go:42: 20:19:24 | es-rollover-autoprov/10-report-spans | starting test step 10-report-spans logger.go:42: 20:19:24 | es-rollover-autoprov/10-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 20:19:32 | es-rollover-autoprov/10-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=03 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-03-job.yaml] logger.go:42: 20:19:32 | es-rollover-autoprov/10-report-spans | running command: [sh -c kubectl apply -f report-span-03-job.yaml -n $NAMESPACE] logger.go:42: 20:19:33 | es-rollover-autoprov/10-report-spans | job.batch/03-report-span created logger.go:42: 20:19:57 | es-rollover-autoprov/10-report-spans | test step completed 10-report-spans logger.go:42: 20:19:57 | es-rollover-autoprov/11-check-indices | starting test step 11-check-indices logger.go:42: 20:19:57 | es-rollover-autoprov/11-check-indices | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob my-jaeger-es-rollover --namespace $NAMESPACE] logger.go:42: 20:19:57 | es-rollover-autoprov/11-check-indices | go: downloading golang.org/x/sys v0.27.0 logger.go:42: 20:19:57 | es-rollover-autoprov/11-check-indices | go: downloading golang.org/x/text v0.20.0 logger.go:42: 20:19:57 | es-rollover-autoprov/11-check-indices | go: downloading golang.org/x/net v0.30.0 logger.go:42: 20:19:57 | es-rollover-autoprov/11-check-indices | go: downloading golang.org/x/term v0.25.0 logger.go:42: 20:19:57 | es-rollover-autoprov/11-check-indices | go: downloading golang.org/x/oauth2 v0.23.0 logger.go:42: 20:19:57 | es-rollover-autoprov/11-check-indices | go: downloading google.golang.org/protobuf v1.35.1 logger.go:42: 20:20:07 | es-rollover-autoprov/11-check-indices | time="2025-02-17T20:20:07Z" level=debug msg="Checking if the my-jaeger-es-rollover CronJob exists" logger.go:42: 20:20:07 | es-rollover-autoprov/11-check-indices | time="2025-02-17T20:20:07Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 20:20:07 | es-rollover-autoprov/11-check-indices | time="2025-02-17T20:20:07Z" level=info msg="Cronjob my-jaeger-es-rollover found successfully" logger.go:42: 20:20:07 | es-rollover-autoprov/11-check-indices | time="2025-02-17T20:20:07Z" level=debug msg="Waiting for the next scheduled job from my-jaeger-es-rollover cronjob" logger.go:42: 20:20:07 | es-rollover-autoprov/11-check-indices | time="2025-02-17T20:20:07Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 20:20:17 | es-rollover-autoprov/11-check-indices | time="2025-02-17T20:20:17Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 20:20:27 | es-rollover-autoprov/11-check-indices | time="2025-02-17T20:20:27Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 20:20:37 | es-rollover-autoprov/11-check-indices | time="2025-02-17T20:20:37Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 20:20:47 | es-rollover-autoprov/11-check-indices | time="2025-02-17T20:20:47Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 20:20:57 | es-rollover-autoprov/11-check-indices | time="2025-02-17T20:20:57Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 20:21:07 | es-rollover-autoprov/11-check-indices | time="2025-02-17T20:21:07Z" level=info msg="Job of owner my-jaeger-es-rollover succeeded after my-jaeger-es-rollover 1m0.049661049s" logger.go:42: 20:21:08 | es-rollover-autoprov/11-check-indices | Job:kuttl-test-present-escargot/05-check-indices created logger.go:42: 20:21:13 | es-rollover-autoprov/11-check-indices | test step completed 11-check-indices logger.go:42: 20:21:13 | es-rollover-autoprov/12-check-indices | starting test step 12-check-indices logger.go:42: 20:21:13 | es-rollover-autoprov/12-check-indices | Job:kuttl-test-present-escargot/06-check-indices created logger.go:42: 20:21:17 | es-rollover-autoprov/12-check-indices | test step completed 12-check-indices logger.go:42: 20:21:17 | es-rollover-autoprov | es-rollover-autoprov events from ns kuttl-test-present-escargot: logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:11 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestpresentescargotmyjaeger-1-84f7c64856 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestpresentescargotmyjaeger-1-84f7chz9p5 replicaset-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentescargotmyjaeger-1-84f7chz9p5 Binding Scheduled Successfully assigned kuttl-test-present-escargot/elasticsearch-cdm-kuttltestpresentescargotmyjaeger-1-84f7chz9p5 to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentescargotmyjaeger-1-84f7chz9p5 AddedInterface Add eth0 [10.129.2.26/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentescargotmyjaeger-1-84f7chz9p5.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:5117d0aee260f74d6ef77c7c71ed073b1ad45b456f58894539b5697bec878310" already present on machine kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentescargotmyjaeger-1-84f7chz9p5.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentescargotmyjaeger-1-84f7chz9p5.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentescargotmyjaeger-1-84f7chz9p5.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d63d612cdca72f64c45960a9b30bcfd0c746daabc17407df92c62694d5bc4bb6" already present on machine kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:11 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentescargotmyjaeger-1-84f7chz9p5.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:11 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestpresentescargotmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestpresentescargotmyjaeger-1-84f7c64856 to 1 deployment-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:12 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentescargotmyjaeger-1-84f7chz9p5.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:21 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpresentescargotmyjaeger-1-84f7chz9p5.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:26 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpresentescargotmyjaeger-1-84f7chz9p5.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:37 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-59bb79f97c to 1 deployment-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal Pod my-jaeger-collector-6985dcb65-sbvk6 Binding Scheduled Successfully assigned kuttl-test-present-escargot/my-jaeger-collector-6985dcb65-sbvk6 to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal Pod my-jaeger-collector-6985dcb65-sbvk6 AddedInterface Add eth0 [10.131.0.19/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal Pod my-jaeger-collector-6985dcb65-sbvk6.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal Pod my-jaeger-collector-6985dcb65-sbvk6.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal Pod my-jaeger-collector-6985dcb65-sbvk6.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-6985dcb65 SuccessfulCreate Created pod: my-jaeger-collector-6985dcb65-sbvk6 replicaset-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-6985dcb65 to 1 deployment-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal Pod my-jaeger-query-59bb79f97c-kn4f2 Binding Scheduled Successfully assigned kuttl-test-present-escargot/my-jaeger-query-59bb79f97c-kn4f2 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal Pod my-jaeger-query-59bb79f97c-kn4f2 AddedInterface Add eth0 [10.128.2.19/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal Pod my-jaeger-query-59bb79f97c-kn4f2.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal Pod my-jaeger-query-59bb79f97c-kn4f2.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal Pod my-jaeger-query-59bb79f97c-kn4f2.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal Pod my-jaeger-query-59bb79f97c-kn4f2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal Pod my-jaeger-query-59bb79f97c-kn4f2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal Pod my-jaeger-query-59bb79f97c-kn4f2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal Pod my-jaeger-query-59bb79f97c-kn4f2.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal Pod my-jaeger-query-59bb79f97c-kn4f2.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal Pod my-jaeger-query-59bb79f97c-kn4f2.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:38 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-59bb79f97c SuccessfulCreate Created pod: my-jaeger-query-59bb79f97c-kn4f2 replicaset-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:46 +0000 UTC Normal Pod my-jaeger-query-59bb79f97c-kn4f2.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:46 +0000 UTC Normal Pod my-jaeger-query-59bb79f97c-kn4f2.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:46 +0000 UTC Normal Pod my-jaeger-query-59bb79f97c-kn4f2.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:46 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-59bb79f97c SuccessfulDelete Deleted pod: my-jaeger-query-59bb79f97c-kn4f2 replicaset-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:46 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-59bb79f97c to 0 from 1 deployment-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:46 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-699fbd55f7 to 1 deployment-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:47 +0000 UTC Normal Pod my-jaeger-query-699fbd55f7-jv2bs Binding Scheduled Successfully assigned kuttl-test-present-escargot/my-jaeger-query-699fbd55f7-jv2bs to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:47 +0000 UTC Normal Pod my-jaeger-query-699fbd55f7-jv2bs AddedInterface Add eth0 [10.128.2.20/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:47 +0000 UTC Normal Pod my-jaeger-query-699fbd55f7-jv2bs.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:47 +0000 UTC Normal Pod my-jaeger-query-699fbd55f7-jv2bs.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:47 +0000 UTC Normal Pod my-jaeger-query-699fbd55f7-jv2bs.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:47 +0000 UTC Normal Pod my-jaeger-query-699fbd55f7-jv2bs.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:47 +0000 UTC Normal Pod my-jaeger-query-699fbd55f7-jv2bs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:47 +0000 UTC Normal Pod my-jaeger-query-699fbd55f7-jv2bs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:47 +0000 UTC Normal Pod my-jaeger-query-699fbd55f7-jv2bs.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:47 +0000 UTC Normal Pod my-jaeger-query-699fbd55f7-jv2bs.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:47 +0000 UTC Normal Pod my-jaeger-query-699fbd55f7-jv2bs.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:47 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-699fbd55f7 SuccessfulCreate Created pod: my-jaeger-query-699fbd55f7-jv2bs replicaset-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:51 +0000 UTC Normal Pod 00-report-span-9stbt Binding Scheduled Successfully assigned kuttl-test-present-escargot/00-report-span-9stbt to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:51 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-9stbt job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:52 +0000 UTC Normal Pod 00-report-span-9stbt AddedInterface Add eth0 [10.131.0.20/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:52 +0000 UTC Normal Pod 00-report-span-9stbt.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:53 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:53 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:53 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:54 +0000 UTC Normal Pod 00-report-span-9stbt.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.219s (2.219s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:54 +0000 UTC Normal Pod 00-report-span-9stbt.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:17:54 +0000 UTC Normal Pod 00-report-span-9stbt.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:08 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-6985dcb65-sbvk6 horizontal-pod-autoscaler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:17 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:18 +0000 UTC Normal Pod 00-check-indices-9bt66 Binding Scheduled Successfully assigned kuttl-test-present-escargot/00-check-indices-9bt66 to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:18 +0000 UTC Normal Pod 00-check-indices-9bt66 AddedInterface Add eth0 [10.131.0.21/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:18 +0000 UTC Normal Pod 00-check-indices-9bt66.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:18 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-9bt66 job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:19 +0000 UTC Normal Pod 00-check-indices-9bt66.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 469ms (469ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:19 +0000 UTC Normal Pod 00-check-indices-9bt66.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:19 +0000 UTC Normal Pod 00-check-indices-9bt66.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:22 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:23 +0000 UTC Normal Pod 01-check-indices-skr2f Binding Scheduled Successfully assigned kuttl-test-present-escargot/01-check-indices-skr2f to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:23 +0000 UTC Normal Pod 01-check-indices-skr2f AddedInterface Add eth0 [10.131.0.22/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:23 +0000 UTC Normal Pod 01-check-indices-skr2f.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:23 +0000 UTC Normal Job.batch 01-check-indices SuccessfulCreate Created pod: 01-check-indices-skr2f job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:23 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-6985dcb65-sbvk6 horizontal-pod-autoscaler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:23 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-6985dcb65-sbvk6 horizontal-pod-autoscaler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:24 +0000 UTC Normal Pod 01-check-indices-skr2f.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 640ms (640ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:24 +0000 UTC Normal Pod 01-check-indices-skr2f.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:24 +0000 UTC Normal Pod 01-check-indices-skr2f.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:27 +0000 UTC Normal Job.batch 01-check-indices Completed Job completed job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:28 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-4tmsj Binding Scheduled Successfully assigned kuttl-test-present-escargot/my-jaeger-es-rollover-create-mapping-4tmsj to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:28 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping SuccessfulCreate Created pod: my-jaeger-es-rollover-create-mapping-4tmsj job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:29 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-4tmsj AddedInterface Add eth0 [10.131.0.23/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:29 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-4tmsj.spec.containers{my-jaeger-es-rollover-create-mapping} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:146ea9aab82c1e7dd871e269a58e0491439b4ce3b98c281e46214fb813cd8194" kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:31 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-4tmsj.spec.containers{my-jaeger-es-rollover-create-mapping} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:146ea9aab82c1e7dd871e269a58e0491439b4ce3b98c281e46214fb813cd8194" in 2.255s (2.255s including waiting). Image size: 104079089 bytes. kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:31 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-4tmsj.spec.containers{my-jaeger-es-rollover-create-mapping} Created Created container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:31 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-4tmsj.spec.containers{my-jaeger-es-rollover-create-mapping} Started Started container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:34 +0000 UTC Normal Pod my-jaeger-collector-6985dcb65-sbvk6.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:34 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-6985dcb65 SuccessfulDelete Deleted pod: my-jaeger-collector-6985dcb65-sbvk6 replicaset-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:34 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-6985dcb65 to 0 from 1 deployment-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:34 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping Completed Job completed job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:34 +0000 UTC Normal Pod my-jaeger-query-699fbd55f7-jv2bs.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:34 +0000 UTC Normal Pod my-jaeger-query-699fbd55f7-jv2bs.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:34 +0000 UTC Normal Pod my-jaeger-query-699fbd55f7-jv2bs.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:34 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-699fbd55f7 SuccessfulDelete Deleted pod: my-jaeger-query-699fbd55f7-jv2bs replicaset-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:34 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-699fbd55f7 to 0 from 1 deployment-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal Pod my-jaeger-collector-64dd56fddd-24npk Binding Scheduled Successfully assigned kuttl-test-present-escargot/my-jaeger-collector-64dd56fddd-24npk to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal Pod my-jaeger-collector-64dd56fddd-24npk AddedInterface Add eth0 [10.131.0.24/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal Pod my-jaeger-collector-64dd56fddd-24npk.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal Pod my-jaeger-collector-64dd56fddd-24npk.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal Pod my-jaeger-collector-64dd56fddd-24npk.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-64dd56fddd SuccessfulCreate Created pod: my-jaeger-collector-64dd56fddd-24npk replicaset-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-64dd56fddd to 1 deployment-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal Pod my-jaeger-query-5f8f76999b-9tppg Binding Scheduled Successfully assigned kuttl-test-present-escargot/my-jaeger-query-5f8f76999b-9tppg to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal Pod my-jaeger-query-5f8f76999b-9tppg AddedInterface Add eth0 [10.128.2.21/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal Pod my-jaeger-query-5f8f76999b-9tppg.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal Pod my-jaeger-query-5f8f76999b-9tppg.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal Pod my-jaeger-query-5f8f76999b-9tppg.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal Pod my-jaeger-query-5f8f76999b-9tppg.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal Pod my-jaeger-query-5f8f76999b-9tppg.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal Pod my-jaeger-query-5f8f76999b-9tppg.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal Pod my-jaeger-query-5f8f76999b-9tppg.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal Pod my-jaeger-query-5f8f76999b-9tppg.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5f8f76999b SuccessfulCreate Created pod: my-jaeger-query-5f8f76999b-9tppg replicaset-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:36 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-5f8f76999b to 1 deployment-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:37 +0000 UTC Normal Pod my-jaeger-query-5f8f76999b-9tppg.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:44 +0000 UTC Normal Pod 02-report-span-brtcm Binding Scheduled Successfully assigned kuttl-test-present-escargot/02-report-span-brtcm to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:44 +0000 UTC Normal Job.batch 02-report-span SuccessfulCreate Created pod: 02-report-span-brtcm job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:45 +0000 UTC Normal Pod 02-report-span-brtcm AddedInterface Add eth0 [10.131.0.25/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:45 +0000 UTC Normal Pod 02-report-span-brtcm.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:47 +0000 UTC Normal Pod 02-report-span-brtcm.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.416s (1.416s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:47 +0000 UTC Normal Pod 02-report-span-brtcm.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:18:47 +0000 UTC Normal Pod 02-report-span-brtcm.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28997059-6nj8m Binding Scheduled Successfully assigned kuttl-test-present-escargot/my-jaeger-es-lookback-28997059-6nj8m to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28997059-6nj8m AddedInterface Add eth0 [10.128.2.22/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28997059-6nj8m.spec.containers{my-jaeger-es-lookback} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:146ea9aab82c1e7dd871e269a58e0491439b4ce3b98c281e46214fb813cd8194" kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28997059 SuccessfulCreate Created pod: my-jaeger-es-lookback-28997059-6nj8m job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28997059 cronjob-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28997059-8d46g Binding Scheduled Successfully assigned kuttl-test-present-escargot/my-jaeger-es-rollover-28997059-8d46g to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28997059-8d46g AddedInterface Add eth0 [10.131.0.26/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28997059-8d46g.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:146ea9aab82c1e7dd871e269a58e0491439b4ce3b98c281e46214fb813cd8194" already present on machine kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28997059-8d46g.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28997059-8d46g.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28997059 SuccessfulCreate Created pod: my-jaeger-es-rollover-28997059-8d46g job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28997059 cronjob-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28997059-6nj8m.spec.containers{my-jaeger-es-lookback} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:146ea9aab82c1e7dd871e269a58e0491439b4ce3b98c281e46214fb813cd8194" in 1.176s (1.176s including waiting). Image size: 104079089 bytes. kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:02 +0000 UTC Normal Pod my-jaeger-es-lookback-28997059-6nj8m.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:02 +0000 UTC Normal Pod my-jaeger-es-lookback-28997059-6nj8m.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28997059 Completed Job completed job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28997059, condition: Complete cronjob-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:04 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28997059 Completed Job completed job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:04 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28997059, condition: Complete cronjob-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:10 +0000 UTC Normal Job.batch 02-report-span Completed Job completed job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:11 +0000 UTC Normal Pod 02-check-indices-fmzmx Binding Scheduled Successfully assigned kuttl-test-present-escargot/02-check-indices-fmzmx to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:11 +0000 UTC Normal Job.batch 02-check-indices SuccessfulCreate Created pod: 02-check-indices-fmzmx job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:12 +0000 UTC Normal Pod 02-check-indices-fmzmx AddedInterface Add eth0 [10.131.0.27/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:12 +0000 UTC Normal Pod 02-check-indices-fmzmx.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:12 +0000 UTC Normal Pod 02-check-indices-fmzmx.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 526ms (526ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:12 +0000 UTC Normal Pod 02-check-indices-fmzmx.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:12 +0000 UTC Normal Pod 02-check-indices-fmzmx.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:15 +0000 UTC Normal Job.batch 02-check-indices Completed Job completed job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:16 +0000 UTC Normal Pod 03-check-indices-66btl Binding Scheduled Successfully assigned kuttl-test-present-escargot/03-check-indices-66btl to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:16 +0000 UTC Normal Job.batch 03-check-indices SuccessfulCreate Created pod: 03-check-indices-66btl job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:17 +0000 UTC Normal Pod 03-check-indices-66btl AddedInterface Add eth0 [10.131.0.28/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:17 +0000 UTC Normal Pod 03-check-indices-66btl.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:17 +0000 UTC Normal Pod 03-check-indices-66btl.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 549ms (549ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:17 +0000 UTC Normal Pod 03-check-indices-66btl.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:17 +0000 UTC Normal Pod 03-check-indices-66btl.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:20 +0000 UTC Normal Job.batch 03-check-indices Completed Job completed job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:20 +0000 UTC Normal Pod 04-check-indices-kbzld Binding Scheduled Successfully assigned kuttl-test-present-escargot/04-check-indices-kbzld to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:20 +0000 UTC Normal Job.batch 04-check-indices SuccessfulCreate Created pod: 04-check-indices-kbzld job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:21 +0000 UTC Normal Pod 04-check-indices-kbzld AddedInterface Add eth0 [10.131.0.29/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:21 +0000 UTC Normal Pod 04-check-indices-kbzld.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:21 +0000 UTC Normal Pod 04-check-indices-kbzld.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 595ms (595ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:21 +0000 UTC Normal Pod 04-check-indices-kbzld.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:22 +0000 UTC Normal Pod 04-check-indices-kbzld.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:23 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-64dd56fddd-24npk horizontal-pod-autoscaler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:24 +0000 UTC Normal Job.batch 04-check-indices Completed Job completed job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:33 +0000 UTC Normal Pod 03-report-span-qm4wg Binding Scheduled Successfully assigned kuttl-test-present-escargot/03-report-span-qm4wg to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:33 +0000 UTC Normal Pod 03-report-span-qm4wg AddedInterface Add eth0 [10.131.0.30/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:33 +0000 UTC Normal Pod 03-report-span-qm4wg.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:33 +0000 UTC Normal Job.batch 03-report-span SuccessfulCreate Created pod: 03-report-span-qm4wg job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:34 +0000 UTC Normal Pod 03-report-span-qm4wg.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 567ms (567ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:34 +0000 UTC Normal Pod 03-report-span-qm4wg.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:34 +0000 UTC Normal Pod 03-report-span-qm4wg.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:38 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-64dd56fddd-24npk horizontal-pod-autoscaler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:38 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-64dd56fddd-24npk horizontal-pod-autoscaler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:19:57 +0000 UTC Normal Job.batch 03-report-span Completed Job completed job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:20:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28997060-x6tdv Binding Scheduled Successfully assigned kuttl-test-present-escargot/my-jaeger-es-lookback-28997060-x6tdv to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:20:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28997060-x6tdv AddedInterface Add eth0 [10.131.0.31/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:20:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28997060-x6tdv.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:146ea9aab82c1e7dd871e269a58e0491439b4ce3b98c281e46214fb813cd8194" already present on machine kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:20:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28997060-x6tdv.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:20:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28997060-x6tdv.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:20:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28997060 SuccessfulCreate Created pod: my-jaeger-es-lookback-28997060-x6tdv job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:20:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28997060 cronjob-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:20:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28997060-n2g8f Binding Scheduled Successfully assigned kuttl-test-present-escargot/my-jaeger-es-rollover-28997060-n2g8f to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:20:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28997060-n2g8f AddedInterface Add eth0 [10.131.0.32/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:20:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28997060-n2g8f.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:146ea9aab82c1e7dd871e269a58e0491439b4ce3b98c281e46214fb813cd8194" already present on machine kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:20:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28997060-n2g8f.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:20:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28997060-n2g8f.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:20:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28997060 SuccessfulCreate Created pod: my-jaeger-es-rollover-28997060-n2g8f job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:20:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28997060 cronjob-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:20:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28997060 Completed Job completed job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:20:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28997060, condition: Complete cronjob-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:20:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28997060 Completed Job completed job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:20:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28997060, condition: Complete cronjob-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28997061-dmz8f Binding Scheduled Successfully assigned kuttl-test-present-escargot/my-jaeger-es-lookback-28997061-dmz8f to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28997061-dmz8f AddedInterface Add eth0 [10.131.0.34/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28997061-dmz8f.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:146ea9aab82c1e7dd871e269a58e0491439b4ce3b98c281e46214fb813cd8194" already present on machine kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28997061-dmz8f.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28997061-dmz8f.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28997061 SuccessfulCreate Created pod: my-jaeger-es-lookback-28997061-dmz8f job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28997061 cronjob-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28997061-5mh9v Binding Scheduled Successfully assigned kuttl-test-present-escargot/my-jaeger-es-rollover-28997061-5mh9v to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28997061-5mh9v AddedInterface Add eth0 [10.131.0.33/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28997061-5mh9v.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:146ea9aab82c1e7dd871e269a58e0491439b4ce3b98c281e46214fb813cd8194" already present on machine kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28997061-5mh9v.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28997061-5mh9v.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28997061 SuccessfulCreate Created pod: my-jaeger-es-rollover-28997061-5mh9v job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28997061 cronjob-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28997061 Completed Job completed job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28997061, condition: Complete cronjob-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28997061 Completed Job completed job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28997061, condition: Complete cronjob-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:08 +0000 UTC Normal Pod 05-check-indices-mwhpd Binding Scheduled Successfully assigned kuttl-test-present-escargot/05-check-indices-mwhpd to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:08 +0000 UTC Normal Pod 05-check-indices-mwhpd AddedInterface Add eth0 [10.131.0.35/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:08 +0000 UTC Normal Pod 05-check-indices-mwhpd.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:08 +0000 UTC Normal Job.batch 05-check-indices SuccessfulCreate Created pod: 05-check-indices-mwhpd job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:09 +0000 UTC Normal Pod 05-check-indices-mwhpd.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 829ms (829ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:09 +0000 UTC Normal Pod 05-check-indices-mwhpd.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:09 +0000 UTC Normal Pod 05-check-indices-mwhpd.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:12 +0000 UTC Normal Job.batch 05-check-indices Completed Job completed job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:13 +0000 UTC Normal Pod 06-check-indices-4lm9x Binding Scheduled Successfully assigned kuttl-test-present-escargot/06-check-indices-4lm9x to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:13 +0000 UTC Normal Pod 06-check-indices-4lm9x AddedInterface Add eth0 [10.131.0.36/23] from ovn-kubernetes multus logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:13 +0000 UTC Normal Pod 06-check-indices-4lm9x.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:13 +0000 UTC Normal Job.batch 06-check-indices SuccessfulCreate Created pod: 06-check-indices-4lm9x job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:14 +0000 UTC Normal Pod 06-check-indices-4lm9x.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 586ms (586ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:14 +0000 UTC Normal Pod 06-check-indices-4lm9x.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:14 +0000 UTC Normal Pod 06-check-indices-4lm9x.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:21:17 | es-rollover-autoprov | 2025-02-17 20:21:16 +0000 UTC Normal Job.batch 06-check-indices Completed Job completed job-controller logger.go:42: 20:21:17 | es-rollover-autoprov | Deleting namespace: kuttl-test-present-escargot === CONT kuttl/harness/es-increasing-replicas logger.go:42: 20:21:25 | es-increasing-replicas | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:21:25 | es-increasing-replicas | Ignoring check-es-nodes.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:21:25 | es-increasing-replicas | Ignoring openshift-check-es-nodes.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:21:25 | es-increasing-replicas | Creating namespace: kuttl-test-champion-oryx logger.go:42: 20:21:25 | es-increasing-replicas/1-install | starting test step 1-install logger.go:42: 20:21:25 | es-increasing-replicas/1-install | Jaeger:kuttl-test-champion-oryx/simple-prod created logger.go:42: 20:22:01 | es-increasing-replicas/1-install | test step completed 1-install logger.go:42: 20:22:01 | es-increasing-replicas/2-install | starting test step 2-install logger.go:42: 20:22:01 | es-increasing-replicas/2-install | Jaeger:kuttl-test-champion-oryx/simple-prod updated logger.go:42: 20:22:20 | es-increasing-replicas/2-install | test step completed 2-install logger.go:42: 20:22:20 | es-increasing-replicas/3-smoke-test | starting test step 3-smoke-test logger.go:42: 20:22:20 | es-increasing-replicas/3-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 20:22:21 | es-increasing-replicas/3-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:22:28 | es-increasing-replicas/3-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 20:22:28 | es-increasing-replicas/3-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 20:22:28 | es-increasing-replicas/3-smoke-test | job.batch/report-span created logger.go:42: 20:22:28 | es-increasing-replicas/3-smoke-test | job.batch/check-span created logger.go:42: 20:22:42 | es-increasing-replicas/3-smoke-test | test step completed 3-smoke-test logger.go:42: 20:22:42 | es-increasing-replicas/4-install | starting test step 4-install logger.go:42: 20:22:42 | es-increasing-replicas/4-install | Jaeger:kuttl-test-champion-oryx/simple-prod updated logger.go:42: 20:22:42 | es-increasing-replicas/4-install | test step completed 4-install logger.go:42: 20:22:42 | es-increasing-replicas/5-check-es-nodes | starting test step 5-check-es-nodes logger.go:42: 20:22:42 | es-increasing-replicas/5-check-es-nodes | running command: [sh -c ./check-es-nodes.sh $NAMESPACE] logger.go:42: 20:22:42 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 20:22:42 | es-increasing-replicas/5-check-es-nodes | false logger.go:42: 20:22:42 | es-increasing-replicas/5-check-es-nodes | Error: no matches found logger.go:42: 20:22:47 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 20:22:47 | es-increasing-replicas/5-check-es-nodes | true logger.go:42: 20:22:47 | es-increasing-replicas/5-check-es-nodes | test step completed 5-check-es-nodes logger.go:42: 20:22:47 | es-increasing-replicas | es-increasing-replicas events from ns kuttl-test-champion-oryx: logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:32 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestchampionoryxsimpleprod-1-7f46c48 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestchampionoryxsimpleprod-1-7f46c4jm527 replicaset-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchampionoryxsimpleprod-1-7f46c4jm527 Binding Scheduled Successfully assigned kuttl-test-champion-oryx/elasticsearch-cdm-kuttltestchampionoryxsimpleprod-1-7f46c4jm527 to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:32 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestchampionoryxsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestchampionoryxsimpleprod-1-7f46c48 to 1 deployment-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchampionoryxsimpleprod-1-7f46c4jm527 AddedInterface Add eth0 [10.129.2.27/23] from ovn-kubernetes multus logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchampionoryxsimpleprod-1-7f46c4jm527.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:5117d0aee260f74d6ef77c7c71ed073b1ad45b456f58894539b5697bec878310" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchampionoryxsimpleprod-1-7f46c4jm527.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchampionoryxsimpleprod-1-7f46c4jm527.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchampionoryxsimpleprod-1-7f46c4jm527.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d63d612cdca72f64c45960a9b30bcfd0c746daabc17407df92c62694d5bc4bb6" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchampionoryxsimpleprod-1-7f46c4jm527.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchampionoryxsimpleprod-1-7f46c4jm527.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:47 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestchampionoryxsimpleprod-1-7f46c4jm527.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:59 +0000 UTC Normal Pod simple-prod-collector-758fbd8999-765kz Binding Scheduled Successfully assigned kuttl-test-champion-oryx/simple-prod-collector-758fbd8999-765kz to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:59 +0000 UTC Normal Pod simple-prod-collector-758fbd8999-765kz AddedInterface Add eth0 [10.128.2.23/23] from ovn-kubernetes multus logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:59 +0000 UTC Normal Pod simple-prod-collector-758fbd8999-765kz.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:59 +0000 UTC Normal Pod simple-prod-collector-758fbd8999-765kz.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:59 +0000 UTC Normal Pod simple-prod-collector-758fbd8999-765kz.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:59 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-758fbd8999 SuccessfulCreate Created pod: simple-prod-collector-758fbd8999-765kz replicaset-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:59 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-758fbd8999 to 1 deployment-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:59 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-p6fg6 Binding Scheduled Successfully assigned kuttl-test-champion-oryx/simple-prod-query-55d4d9b5f6-p6fg6 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:59 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-p6fg6 AddedInterface Add eth0 [10.128.2.24/23] from ovn-kubernetes multus logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:59 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-p6fg6.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:59 +0000 UTC Normal ReplicaSet.apps simple-prod-query-55d4d9b5f6 SuccessfulCreate Created pod: simple-prod-query-55d4d9b5f6-p6fg6 replicaset-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:21:59 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-55d4d9b5f6 to 1 deployment-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:00 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-p6fg6.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:00 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-p6fg6.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:00 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-p6fg6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:00 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-p6fg6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:00 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-p6fg6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:00 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-p6fg6.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:00 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-p6fg6.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:00 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-p6fg6.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:06 +0000 UTC Normal Pod simple-prod-collector-758fbd8999-t6tpq Binding Scheduled Successfully assigned kuttl-test-champion-oryx/simple-prod-collector-758fbd8999-t6tpq to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:06 +0000 UTC Normal Pod simple-prod-collector-758fbd8999-t6tpq AddedInterface Add eth0 [10.131.0.37/23] from ovn-kubernetes multus logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:06 +0000 UTC Normal Pod simple-prod-collector-758fbd8999-t6tpq.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:06 +0000 UTC Normal Pod simple-prod-collector-758fbd8999-t6tpq.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:06 +0000 UTC Normal Pod simple-prod-collector-758fbd8999-t6tpq.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:06 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-758fbd8999 SuccessfulCreate Created pod: simple-prod-collector-758fbd8999-t6tpq replicaset-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:06 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-758fbd8999 to 2 from 1 deployment-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:06 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-zd9th Binding Scheduled Successfully assigned kuttl-test-champion-oryx/simple-prod-query-55d4d9b5f6-zd9th to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:06 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-zd9th AddedInterface Add eth0 [10.131.0.38/23] from ovn-kubernetes multus logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:06 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-zd9th.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:06 +0000 UTC Normal ReplicaSet.apps simple-prod-query-55d4d9b5f6 SuccessfulCreate Created pod: simple-prod-query-55d4d9b5f6-zd9th replicaset-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:06 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-55d4d9b5f6 to 2 from 1 deployment-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:09 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-zd9th.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" in 2.708s (2.708s including waiting). Image size: 142020742 bytes. kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:09 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-zd9th.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:09 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-zd9th.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:09 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-zd9th.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:13 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-zd9th.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" in 3.825s (3.825s including waiting). Image size: 438322369 bytes. kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:13 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-zd9th.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:13 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-zd9th.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:13 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-zd9th.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:18 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-zd9th.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" in 5.23s (5.23s including waiting). Image size: 112614125 bytes. kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:18 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-zd9th.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:18 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-zd9th.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:23 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-p6fg6.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:23 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-p6fg6.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:23 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-p6fg6.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:23 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-zd9th.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:23 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-zd9th.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:23 +0000 UTC Normal Pod simple-prod-query-55d4d9b5f6-zd9th.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:23 +0000 UTC Normal ReplicaSet.apps simple-prod-query-55d4d9b5f6 SuccessfulDelete Deleted pod: simple-prod-query-55d4d9b5f6-p6fg6 replicaset-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:23 +0000 UTC Normal ReplicaSet.apps simple-prod-query-55d4d9b5f6 SuccessfulDelete Deleted pod: simple-prod-query-55d4d9b5f6-zd9th replicaset-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:23 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-55d4d9b5f6 to 0 from 2 deployment-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:24 +0000 UTC Normal Pod simple-prod-query-568858db74-27rcb Binding Scheduled Successfully assigned kuttl-test-champion-oryx/simple-prod-query-568858db74-27rcb to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:24 +0000 UTC Normal Pod simple-prod-query-568858db74-6xkt2 Binding Scheduled Successfully assigned kuttl-test-champion-oryx/simple-prod-query-568858db74-6xkt2 to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:24 +0000 UTC Normal ReplicaSet.apps simple-prod-query-568858db74 SuccessfulCreate Created pod: simple-prod-query-568858db74-27rcb replicaset-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:24 +0000 UTC Normal ReplicaSet.apps simple-prod-query-568858db74 SuccessfulCreate Created pod: simple-prod-query-568858db74-6xkt2 replicaset-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:24 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-568858db74 to 2 deployment-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-27rcb AddedInterface Add eth0 [10.128.2.25/23] from ovn-kubernetes multus logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-27rcb.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-27rcb.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-27rcb.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-27rcb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-27rcb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-27rcb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-27rcb.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-27rcb.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-27rcb.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-6xkt2 AddedInterface Add eth0 [10.131.0.39/23] from ovn-kubernetes multus logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-6xkt2.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-6xkt2.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-6xkt2.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-6xkt2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-6xkt2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-6xkt2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-6xkt2.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-6xkt2.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:25 +0000 UTC Normal Pod simple-prod-query-568858db74-6xkt2.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:28 +0000 UTC Normal Pod check-span-w77wd Binding Scheduled Successfully assigned kuttl-test-champion-oryx/check-span-w77wd to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:28 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-w77wd job-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:28 +0000 UTC Normal Pod report-span-jkh2g Binding Scheduled Successfully assigned kuttl-test-champion-oryx/report-span-jkh2g to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:28 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-jkh2g job-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:29 +0000 UTC Normal Pod check-span-w77wd AddedInterface Add eth0 [10.128.2.26/23] from ovn-kubernetes multus logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:29 +0000 UTC Normal Pod check-span-w77wd.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:29 +0000 UTC Normal Pod report-span-jkh2g AddedInterface Add eth0 [10.131.0.40/23] from ovn-kubernetes multus logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:29 +0000 UTC Normal Pod report-span-jkh2g.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:30 +0000 UTC Normal Pod report-span-jkh2g.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 919ms (919ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:30 +0000 UTC Normal Pod report-span-jkh2g.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:30 +0000 UTC Normal Pod report-span-jkh2g.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:31 +0000 UTC Normal Pod check-span-w77wd.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.653s (1.653s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:31 +0000 UTC Normal Pod check-span-w77wd.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:31 +0000 UTC Normal Pod check-span-w77wd.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:41 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:43 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestchampionoryxsimpleprod-2-68994ccbb SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestchampionoryxsimpleprod-2-68994ccfjwd replicaset-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchampionoryxsimpleprod-2-68994ccfjwd Binding Scheduled Successfully assigned kuttl-test-champion-oryx/elasticsearch-cdm-kuttltestchampionoryxsimpleprod-2-68994ccfjwd to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:43 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestchampionoryxsimpleprod-2 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestchampionoryxsimpleprod-2-68994ccbb to 1 deployment-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:43 +0000 UTC Normal Pod simple-prod-collector-758fbd8999-765kz.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:43 +0000 UTC Normal Pod simple-prod-collector-758fbd8999-t6tpq.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:43 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-758fbd8999 SuccessfulDelete Deleted pod: simple-prod-collector-758fbd8999-765kz replicaset-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:43 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-758fbd8999 SuccessfulDelete Deleted pod: simple-prod-collector-758fbd8999-t6tpq replicaset-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:43 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled down replica set simple-prod-collector-758fbd8999 to 0 from 2 deployment-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:43 +0000 UTC Normal Pod simple-prod-query-568858db74-27rcb.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:43 +0000 UTC Normal Pod simple-prod-query-568858db74-27rcb.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:43 +0000 UTC Normal Pod simple-prod-query-568858db74-27rcb.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:43 +0000 UTC Normal Pod simple-prod-query-568858db74-6xkt2.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:43 +0000 UTC Normal Pod simple-prod-query-568858db74-6xkt2.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:43 +0000 UTC Normal Pod simple-prod-query-568858db74-6xkt2.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:43 +0000 UTC Normal ReplicaSet.apps simple-prod-query-568858db74 SuccessfulDelete Deleted pod: simple-prod-query-568858db74-6xkt2 replicaset-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:43 +0000 UTC Normal ReplicaSet.apps simple-prod-query-568858db74 SuccessfulDelete Deleted pod: simple-prod-query-568858db74-27rcb replicaset-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:43 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-568858db74 to 0 from 2 deployment-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchampionoryxsimpleprod-2-68994ccfjwd AddedInterface Add eth0 [10.131.0.41/23] from ovn-kubernetes multus logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchampionoryxsimpleprod-2-68994ccfjwd.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:5117d0aee260f74d6ef77c7c71ed073b1ad45b456f58894539b5697bec878310" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchampionoryxsimpleprod-2-68994ccfjwd.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchampionoryxsimpleprod-2-68994ccfjwd.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchampionoryxsimpleprod-2-68994ccfjwd.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d63d612cdca72f64c45960a9b30bcfd0c746daabc17407df92c62694d5bc4bb6" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchampionoryxsimpleprod-2-68994ccfjwd.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestchampionoryxsimpleprod-2-68994ccfjwd.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-collector-8696d66948-hn78m Binding Scheduled Successfully assigned kuttl-test-champion-oryx/simple-prod-collector-8696d66948-hn78m to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-collector-8696d66948-hn78m AddedInterface Add eth0 [10.128.2.28/23] from ovn-kubernetes multus logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-collector-8696d66948-hn78m.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-collector-8696d66948-hn78m.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-collector-8696d66948-hn78m.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-collector-8696d66948-pj4rl Binding Scheduled Successfully assigned kuttl-test-champion-oryx/simple-prod-collector-8696d66948-pj4rl to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-collector-8696d66948-pj4rl AddedInterface Add eth0 [10.131.0.43/23] from ovn-kubernetes multus logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-collector-8696d66948-pj4rl.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-8696d66948 SuccessfulCreate Created pod: simple-prod-collector-8696d66948-hn78m replicaset-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-8696d66948 SuccessfulCreate Created pod: simple-prod-collector-8696d66948-pj4rl replicaset-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-8696d66948 to 2 deployment-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-jxqqx Binding Scheduled Successfully assigned kuttl-test-champion-oryx/simple-prod-query-fdc5c8497-jxqqx to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-jxqqx AddedInterface Add eth0 [10.131.0.42/23] from ovn-kubernetes multus logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-jxqqx.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-kk7cp Binding Scheduled Successfully assigned kuttl-test-champion-oryx/simple-prod-query-fdc5c8497-kk7cp to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-kk7cp AddedInterface Add eth0 [10.128.2.27/23] from ovn-kubernetes multus logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-kk7cp.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-kk7cp.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-kk7cp.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-kk7cp.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-kk7cp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-kk7cp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-kk7cp.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-kk7cp.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal ReplicaSet.apps simple-prod-query-fdc5c8497 SuccessfulCreate Created pod: simple-prod-query-fdc5c8497-kk7cp replicaset-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal ReplicaSet.apps simple-prod-query-fdc5c8497 SuccessfulCreate Created pod: simple-prod-query-fdc5c8497-jxqqx replicaset-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:45 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-fdc5c8497 to 2 deployment-controller logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:46 +0000 UTC Normal Pod simple-prod-collector-8696d66948-pj4rl.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:46 +0000 UTC Normal Pod simple-prod-collector-8696d66948-pj4rl.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:46 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-jxqqx.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:46 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-jxqqx.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:46 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-jxqqx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:46 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-jxqqx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:46 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-jxqqx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:46 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-jxqqx.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:46 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-jxqqx.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:46 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-jxqqx.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:22:47 | es-increasing-replicas | 2025-02-17 20:22:46 +0000 UTC Normal Pod simple-prod-query-fdc5c8497-kk7cp.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:22:47 | es-increasing-replicas | Deleting namespace: kuttl-test-champion-oryx === CONT kuttl/harness/es-index-cleaner-autoprov logger.go:42: 20:23:23 | es-index-cleaner-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:23:23 | es-index-cleaner-autoprov | Creating namespace: kuttl-test-electric-krill logger.go:42: 20:23:23 | es-index-cleaner-autoprov/1-install | starting test step 1-install logger.go:42: 20:23:23 | es-index-cleaner-autoprov/1-install | Jaeger:kuttl-test-electric-krill/test-es-index-cleaner-with-prefix created logger.go:42: 20:23:58 | es-index-cleaner-autoprov/1-install | test step completed 1-install logger.go:42: 20:23:58 | es-index-cleaner-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 20:23:58 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE test-es-index-cleaner-with-prefix /dev/null] logger.go:42: 20:24:00 | es-index-cleaner-autoprov/2-report-spans | Warning: resource jaegers/test-es-index-cleaner-with-prefix is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:24:08 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c DAYS=5 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 20:24:09 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 20:24:09 | es-index-cleaner-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 20:24:48 | es-index-cleaner-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 20:24:48 | es-index-cleaner-autoprov/3-install | starting test step 3-install logger.go:42: 20:24:48 | es-index-cleaner-autoprov/3-install | Jaeger:kuttl-test-electric-krill/test-es-index-cleaner-with-prefix updated logger.go:42: 20:24:48 | es-index-cleaner-autoprov/3-install | test step completed 3-install logger.go:42: 20:24:48 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | starting test step 4-wait-es-index-cleaner logger.go:42: 20:24:48 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob test-es-index-cleaner-with-prefix-es-index-cleaner --namespace $NAMESPACE] logger.go:42: 20:24:49 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2025-02-17T20:24:49Z" level=debug msg="Checking if the test-es-index-cleaner-with-prefix-es-index-cleaner CronJob exists" logger.go:42: 20:24:49 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2025-02-17T20:24:49Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 20:24:49 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2025-02-17T20:24:49Z" level=warning msg="The BatchV1/Cronjob test-es-index-cleaner-with-prefix-es-index-cleaner was not found" logger.go:42: 20:24:49 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2025-02-17T20:24:49Z" level=debug msg="Found BatchV/Cronjobs:" logger.go:42: 20:24:59 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2025-02-17T20:24:59Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 20:24:59 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2025-02-17T20:24:59Z" level=info msg="Cronjob test-es-index-cleaner-with-prefix-es-index-cleaner found successfully" logger.go:42: 20:24:59 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2025-02-17T20:24:59Z" level=debug msg="Waiting for the next scheduled job from test-es-index-cleaner-with-prefix-es-index-cleaner cronjob" logger.go:42: 20:24:59 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2025-02-17T20:24:59Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 20:25:09 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2025-02-17T20:25:09Z" level=info msg="Job of owner test-es-index-cleaner-with-prefix-es-index-cleaner succeeded after test-es-index-cleaner-with-prefix-es-index-cleaner 10.036126541s" logger.go:42: 20:25:09 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | test step completed 4-wait-es-index-cleaner logger.go:42: 20:25:09 | es-index-cleaner-autoprov/5-install | starting test step 5-install logger.go:42: 20:25:09 | es-index-cleaner-autoprov/5-install | Jaeger:kuttl-test-electric-krill/test-es-index-cleaner-with-prefix updated logger.go:42: 20:25:09 | es-index-cleaner-autoprov/5-install | test step completed 5-install logger.go:42: 20:25:09 | es-index-cleaner-autoprov/6-check-indices | starting test step 6-check-indices logger.go:42: 20:25:09 | es-index-cleaner-autoprov/6-check-indices | Job:kuttl-test-electric-krill/00-check-indices created logger.go:42: 20:25:13 | es-index-cleaner-autoprov/6-check-indices | test step completed 6-check-indices logger.go:42: 20:25:13 | es-index-cleaner-autoprov | es-index-cleaner-autoprov events from ns kuttl-test-electric-krill: logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:30 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestelectrickrilltestesindexcle-1-5c98d88f64 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestelectrickrilltestesindexcle-1-5z9kfv replicaset-controller logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:30 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestelectrickrilltestesindexcle-1-5z9kfv Binding Scheduled Successfully assigned kuttl-test-electric-krill/elasticsearch-cdm-kuttltestelectrickrilltestesindexcle-1-5z9kfv to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:30 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestelectrickrilltestesindexcle-1-5z9kfv AddedInterface Add eth0 [10.129.2.28/23] from ovn-kubernetes multus logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:30 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestelectrickrilltestesindexcle-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestelectrickrilltestesindexcle-1-5c98d88f64 to 1 deployment-controller logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestelectrickrilltestesindexcle-1-5z9kfv.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:5117d0aee260f74d6ef77c7c71ed073b1ad45b456f58894539b5697bec878310" already present on machine kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestelectrickrilltestesindexcle-1-5z9kfv.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestelectrickrilltestesindexcle-1-5z9kfv.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestelectrickrilltestesindexcle-1-5z9kfv.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d63d612cdca72f64c45960a9b30bcfd0c746daabc17407df92c62694d5bc4bb6" already present on machine kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestelectrickrilltestesindexcle-1-5z9kfv.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestelectrickrilltestesindexcle-1-5z9kfv.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:45 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestelectrickrilltestesindexcle-1-5z9kfv.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:56 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-77c5c5bfbc-n2vj8 Binding Scheduled Successfully assigned kuttl-test-electric-krill/test-es-index-cleaner-with-prefix-collector-77c5c5bfbc-n2vj8 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:56 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-collector-77c5c5bfbc SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-collector-77c5c5bfbc-n2vj8 replicaset-controller logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:56 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-collector ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-collector-77c5c5bfbc to 1 deployment-controller logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:56 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-84fcbfb66f SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-84fcbfb66f-kzz2s replicaset-controller logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:56 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-84fcbfb66f to 1 deployment-controller logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-77c5c5bfbc-n2vj8 AddedInterface Add eth0 [10.128.2.29/23] from ovn-kubernetes multus logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-77c5c5bfbc-n2vj8.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-77c5c5bfbc-n2vj8.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-77c5c5bfbc-n2vj8.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-84fcbfb66f-kzz2s Binding Scheduled Successfully assigned kuttl-test-electric-krill/test-es-index-cleaner-with-prefix-query-84fcbfb66f-kzz2s to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-84fcbfb66f-kzz2s AddedInterface Add eth0 [10.131.0.44/23] from ovn-kubernetes multus logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-84fcbfb66f-kzz2s.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-84fcbfb66f-kzz2s.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-84fcbfb66f-kzz2s.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-84fcbfb66f-kzz2s.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-84fcbfb66f-kzz2s.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-84fcbfb66f-kzz2s.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-84fcbfb66f-kzz2s.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-84fcbfb66f-kzz2s.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:23:57 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-84fcbfb66f-kzz2s.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:04 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-84fcbfb66f-kzz2s.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:04 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-84fcbfb66f-kzz2s.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:04 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-84fcbfb66f-kzz2s.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:04 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-84fcbfb66f SuccessfulDelete Deleted pod: test-es-index-cleaner-with-prefix-query-84fcbfb66f-kzz2s replicaset-controller logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:04 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled down replica set test-es-index-cleaner-with-prefix-query-84fcbfb66f to 0 from 1 deployment-controller logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:05 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ffb9dd77-f6hcs Binding Scheduled Successfully assigned kuttl-test-electric-krill/test-es-index-cleaner-with-prefix-query-77ffb9dd77-f6hcs to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:05 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ffb9dd77-f6hcs AddedInterface Add eth0 [10.131.0.45/23] from ovn-kubernetes multus logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:05 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ffb9dd77-f6hcs.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:05 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-77ffb9dd77 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-77ffb9dd77-f6hcs replicaset-controller logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:05 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-77ffb9dd77 to 1 deployment-controller logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:06 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ffb9dd77-f6hcs.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:06 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ffb9dd77-f6hcs.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:06 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ffb9dd77-f6hcs.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:06 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ffb9dd77-f6hcs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:06 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ffb9dd77-f6hcs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:06 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ffb9dd77-f6hcs.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:06 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ffb9dd77-f6hcs.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:06 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-77ffb9dd77-f6hcs.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:09 +0000 UTC Normal Pod 00-report-span-mvh5q Binding Scheduled Successfully assigned kuttl-test-electric-krill/00-report-span-mvh5q to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:09 +0000 UTC Normal Pod 00-report-span-mvh5q AddedInterface Add eth0 [10.128.2.30/23] from ovn-kubernetes multus logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:09 +0000 UTC Normal Pod 00-report-span-mvh5q.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:09 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-mvh5q job-controller logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:10 +0000 UTC Normal Pod 00-report-span-mvh5q.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 743ms (743ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:10 +0000 UTC Normal Pod 00-report-span-mvh5q.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:10 +0000 UTC Normal Pod 00-report-span-mvh5q.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:12 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:12 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:12 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:42 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-77c5c5bfbc-n2vj8 horizontal-pod-autoscaler logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:42 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-77c5c5bfbc-n2vj8 horizontal-pod-autoscaler logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:42 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-77c5c5bfbc-n2vj8 horizontal-pod-autoscaler logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:24:48 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:25:00 +0000 UTC Normal Job.batch test-es-index-cleaner-with-prefix-es-index-cleaner-28997065 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-es-index-cleaner-2899706fwv7m job-controller logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:25:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2899706fwv7m Binding Scheduled Successfully assigned kuttl-test-electric-krill/test-es-index-cleaner-with-prefix-es-index-cleaner-2899706fwv7m to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:25:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2899706fwv7m AddedInterface Add eth0 [10.128.2.31/23] from ovn-kubernetes multus logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:25:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2899706fwv7m.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:3d41eed6b25f86a907a5935bd96ed634148af2b8869fd92109448b255c2d17c0" kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:25:00 +0000 UTC Normal CronJob.batch test-es-index-cleaner-with-prefix-es-index-cleaner SuccessfulCreate Created job test-es-index-cleaner-with-prefix-es-index-cleaner-28997065 cronjob-controller logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:25:03 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2899706fwv7m.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:3d41eed6b25f86a907a5935bd96ed634148af2b8869fd92109448b255c2d17c0" in 2.458s (2.458s including waiting). Image size: 103733580 bytes. kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:25:03 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2899706fwv7m.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Created Created container test-es-index-cleaner-with-prefix-es-index-cleaner kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:25:03 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2899706fwv7m.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Started Started container test-es-index-cleaner-with-prefix-es-index-cleaner kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:25:06 +0000 UTC Normal Job.batch test-es-index-cleaner-with-prefix-es-index-cleaner-28997065 Completed Job completed job-controller logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:25:06 +0000 UTC Normal CronJob.batch test-es-index-cleaner-with-prefix-es-index-cleaner SawCompletedJob Saw completed job: test-es-index-cleaner-with-prefix-es-index-cleaner-28997065, condition: Complete cronjob-controller logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:25:09 +0000 UTC Normal Pod 00-check-indices-2p9z6 Binding Scheduled Successfully assigned kuttl-test-electric-krill/00-check-indices-2p9z6 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:25:09 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-2p9z6 job-controller logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:25:10 +0000 UTC Normal Pod 00-check-indices-2p9z6 AddedInterface Add eth0 [10.128.2.32/23] from ovn-kubernetes multus logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:25:10 +0000 UTC Normal Pod 00-check-indices-2p9z6.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:25:11 +0000 UTC Normal Pod 00-check-indices-2p9z6.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 964ms (964ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:25:11 +0000 UTC Normal Pod 00-check-indices-2p9z6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:25:11 +0000 UTC Normal Pod 00-check-indices-2p9z6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:25:13 | es-index-cleaner-autoprov | 2025-02-17 20:25:13 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 20:25:13 | es-index-cleaner-autoprov | Deleting namespace: kuttl-test-electric-krill === CONT kuttl/harness/es-from-aio-to-production logger.go:42: 20:25:20 | es-from-aio-to-production | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:25:20 | es-from-aio-to-production | Creating namespace: kuttl-test-willing-redfish logger.go:42: 20:25:20 | es-from-aio-to-production/0-install | starting test step 0-install logger.go:42: 20:25:20 | es-from-aio-to-production/0-install | Jaeger:kuttl-test-willing-redfish/my-jaeger created logger.go:42: 20:25:28 | es-from-aio-to-production/0-install | test step completed 0-install logger.go:42: 20:25:28 | es-from-aio-to-production/1-smoke-test | starting test step 1-smoke-test logger.go:42: 20:25:28 | es-from-aio-to-production/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 20:25:30 | es-from-aio-to-production/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:25:36 | es-from-aio-to-production/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 20:25:37 | es-from-aio-to-production/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 20:25:37 | es-from-aio-to-production/1-smoke-test | job.batch/report-span created logger.go:42: 20:25:37 | es-from-aio-to-production/1-smoke-test | job.batch/check-span created logger.go:42: 20:25:51 | es-from-aio-to-production/1-smoke-test | test step completed 1-smoke-test logger.go:42: 20:25:51 | es-from-aio-to-production/3-install | starting test step 3-install logger.go:42: 20:25:51 | es-from-aio-to-production/3-install | Jaeger:kuttl-test-willing-redfish/my-jaeger updated logger.go:42: 20:26:25 | es-from-aio-to-production/3-install | test step completed 3-install logger.go:42: 20:26:25 | es-from-aio-to-production/4-smoke-test | starting test step 4-smoke-test logger.go:42: 20:26:25 | es-from-aio-to-production/4-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 20:26:33 | es-from-aio-to-production/4-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 20:26:33 | es-from-aio-to-production/4-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 20:26:33 | es-from-aio-to-production/4-smoke-test | job.batch/report-span unchanged logger.go:42: 20:26:33 | es-from-aio-to-production/4-smoke-test | job.batch/check-span unchanged logger.go:42: 20:26:33 | es-from-aio-to-production/4-smoke-test | test step completed 4-smoke-test logger.go:42: 20:26:33 | es-from-aio-to-production | es-from-aio-to-production events from ns kuttl-test-willing-redfish: logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:24 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-68cbd5c65 to 1 deployment-controller logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:25 +0000 UTC Normal Pod my-jaeger-68cbd5c65-4tvf2 Binding Scheduled Successfully assigned kuttl-test-willing-redfish/my-jaeger-68cbd5c65-4tvf2 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:25 +0000 UTC Normal Pod my-jaeger-68cbd5c65-4tvf2 AddedInterface Add eth0 [10.128.2.33/23] from ovn-kubernetes multus logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:25 +0000 UTC Normal ReplicaSet.apps my-jaeger-68cbd5c65 SuccessfulCreate Created pod: my-jaeger-68cbd5c65-4tvf2 replicaset-controller logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:26 +0000 UTC Normal Pod my-jaeger-68cbd5c65-4tvf2.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:27 +0000 UTC Normal Pod my-jaeger-68cbd5c65-4tvf2.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" in 1.903s (1.903s including waiting). Image size: 144304495 bytes. kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:27 +0000 UTC Normal Pod my-jaeger-68cbd5c65-4tvf2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:27 +0000 UTC Normal Pod my-jaeger-68cbd5c65-4tvf2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:27 +0000 UTC Normal Pod my-jaeger-68cbd5c65-4tvf2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:28 +0000 UTC Normal Pod my-jaeger-68cbd5c65-4tvf2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:28 +0000 UTC Normal Pod my-jaeger-68cbd5c65-4tvf2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:33 +0000 UTC Normal Pod my-jaeger-68cbd5c65-4tvf2.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:33 +0000 UTC Normal Pod my-jaeger-68cbd5c65-4tvf2.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:33 +0000 UTC Normal ReplicaSet.apps my-jaeger-68cbd5c65 SuccessfulDelete Deleted pod: my-jaeger-68cbd5c65-4tvf2 replicaset-controller logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:33 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-68cbd5c65 to 0 from 1 deployment-controller logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:34 +0000 UTC Normal Pod my-jaeger-64dd9b69f4-9dgsv Binding Scheduled Successfully assigned kuttl-test-willing-redfish/my-jaeger-64dd9b69f4-9dgsv to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:34 +0000 UTC Normal ReplicaSet.apps my-jaeger-64dd9b69f4 SuccessfulCreate Created pod: my-jaeger-64dd9b69f4-9dgsv replicaset-controller logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:34 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-64dd9b69f4 to 1 deployment-controller logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:35 +0000 UTC Normal Pod my-jaeger-64dd9b69f4-9dgsv AddedInterface Add eth0 [10.128.2.34/23] from ovn-kubernetes multus logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:35 +0000 UTC Normal Pod my-jaeger-64dd9b69f4-9dgsv.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:35 +0000 UTC Normal Pod my-jaeger-64dd9b69f4-9dgsv.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:35 +0000 UTC Normal Pod my-jaeger-64dd9b69f4-9dgsv.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:35 +0000 UTC Normal Pod my-jaeger-64dd9b69f4-9dgsv.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:35 +0000 UTC Normal Pod my-jaeger-64dd9b69f4-9dgsv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:35 +0000 UTC Normal Pod my-jaeger-64dd9b69f4-9dgsv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:37 +0000 UTC Normal Pod check-span-hxsnk Binding Scheduled Successfully assigned kuttl-test-willing-redfish/check-span-hxsnk to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:37 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-hxsnk job-controller logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:37 +0000 UTC Normal Pod report-span-vwmj2 Binding Scheduled Successfully assigned kuttl-test-willing-redfish/report-span-vwmj2 to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:37 +0000 UTC Normal Pod report-span-vwmj2 AddedInterface Add eth0 [10.129.2.29/23] from ovn-kubernetes multus logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:37 +0000 UTC Normal Pod report-span-vwmj2.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:37 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-vwmj2 job-controller logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:38 +0000 UTC Normal Pod check-span-hxsnk AddedInterface Add eth0 [10.129.2.30/23] from ovn-kubernetes multus logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:38 +0000 UTC Normal Pod check-span-hxsnk.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:39 +0000 UTC Normal Pod check-span-hxsnk.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.573s (1.573s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:39 +0000 UTC Normal Pod check-span-hxsnk.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:39 +0000 UTC Normal Pod check-span-hxsnk.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:39 +0000 UTC Normal Pod report-span-vwmj2.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.655s (1.655s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:39 +0000 UTC Normal Pod report-span-vwmj2.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:39 +0000 UTC Normal Pod report-span-vwmj2.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:50 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:55 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestwillingredfishmyjaeger-1-55d8fd85c8 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestwillingredfishmyjaeger-1-55d8fd9tx6q replicaset-controller logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwillingredfishmyjaeger-1-55d8fd9tx6q Binding Scheduled Successfully assigned kuttl-test-willing-redfish/elasticsearch-cdm-kuttltestwillingredfishmyjaeger-1-55d8fd9tx6q to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:55 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestwillingredfishmyjaeger-1-55d8fd9tx6q FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:55 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestwillingredfishmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestwillingredfishmyjaeger-1-55d8fd85c8 to 1 deployment-controller logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwillingredfishmyjaeger-1-55d8fd9tx6q AddedInterface Add eth0 [10.129.2.31/23] from ovn-kubernetes multus logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwillingredfishmyjaeger-1-55d8fd9tx6q.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:5117d0aee260f74d6ef77c7c71ed073b1ad45b456f58894539b5697bec878310" already present on machine kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwillingredfishmyjaeger-1-55d8fd9tx6q.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwillingredfishmyjaeger-1-55d8fd9tx6q.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwillingredfishmyjaeger-1-55d8fd9tx6q.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d63d612cdca72f64c45960a9b30bcfd0c746daabc17407df92c62694d5bc4bb6" already present on machine kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwillingredfishmyjaeger-1-55d8fd9tx6q.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:25:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwillingredfishmyjaeger-1-55d8fd9tx6q.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:06 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestwillingredfishmyjaeger-1-55d8fd9tx6q.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:11 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestwillingredfishmyjaeger-1-55d8fd9tx6q.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:12 +0000 UTC Normal Job.batch report-span Completed Job completed job-controller logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:22 +0000 UTC Normal Pod my-jaeger-64dd9b69f4-9dgsv.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:22 +0000 UTC Normal Pod my-jaeger-64dd9b69f4-9dgsv.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:22 +0000 UTC Normal Pod my-jaeger-collector-59bb9c55c-w8cjq Binding Scheduled Successfully assigned kuttl-test-willing-redfish/my-jaeger-collector-59bb9c55c-w8cjq to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:22 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-59bb9c55c SuccessfulCreate Created pod: my-jaeger-collector-59bb9c55c-w8cjq replicaset-controller logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:22 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-59bb9c55c to 1 deployment-controller logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:22 +0000 UTC Normal Pod my-jaeger-query-79f6f79c74-nc5f7 Binding Scheduled Successfully assigned kuttl-test-willing-redfish/my-jaeger-query-79f6f79c74-nc5f7 to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:22 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-79f6f79c74 SuccessfulCreate Created pod: my-jaeger-query-79f6f79c74-nc5f7 replicaset-controller logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:22 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-79f6f79c74 to 1 deployment-controller logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:23 +0000 UTC Normal Pod my-jaeger-collector-59bb9c55c-w8cjq AddedInterface Add eth0 [10.131.0.46/23] from ovn-kubernetes multus logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:23 +0000 UTC Normal Pod my-jaeger-collector-59bb9c55c-w8cjq.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:23 +0000 UTC Normal Pod my-jaeger-collector-59bb9c55c-w8cjq.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:23 +0000 UTC Normal Pod my-jaeger-collector-59bb9c55c-w8cjq.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:23 +0000 UTC Normal Pod my-jaeger-query-79f6f79c74-nc5f7 AddedInterface Add eth0 [10.131.0.47/23] from ovn-kubernetes multus logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:23 +0000 UTC Normal Pod my-jaeger-query-79f6f79c74-nc5f7.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:23 +0000 UTC Normal Pod my-jaeger-query-79f6f79c74-nc5f7.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:23 +0000 UTC Normal Pod my-jaeger-query-79f6f79c74-nc5f7.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:23 +0000 UTC Normal Pod my-jaeger-query-79f6f79c74-nc5f7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:23 +0000 UTC Normal Pod my-jaeger-query-79f6f79c74-nc5f7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:23 +0000 UTC Normal Pod my-jaeger-query-79f6f79c74-nc5f7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:23 +0000 UTC Normal Pod my-jaeger-query-79f6f79c74-nc5f7.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:23 +0000 UTC Normal Pod my-jaeger-query-79f6f79c74-nc5f7.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | 2025-02-17 20:26:23 +0000 UTC Normal Pod my-jaeger-query-79f6f79c74-nc5f7.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:26:33 | es-from-aio-to-production | Deleting namespace: kuttl-test-willing-redfish === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (704.29s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.63s) --- PASS: kuttl/harness/es-multiinstance (113.86s) --- PASS: kuttl/harness/es-simple-prod (6.51s) --- PASS: kuttl/harness/es-rollover-autoprov (261.76s) --- PASS: kuttl/harness/es-increasing-replicas (118.12s) --- PASS: kuttl/harness/es-index-cleaner-autoprov (116.68s) --- PASS: kuttl/harness/es-from-aio-to-production (80.57s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name elasticsearch --report --output /logs/artifacts/elasticsearch.xml ./artifacts/kuttl-report.xml time="2025-02-17T20:26:42Z" level=debug msg="Setting a new name for the test suites" time="2025-02-17T20:26:42Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-02-17T20:26:42Z" level=debug msg="normalizing test case names" time="2025-02-17T20:26:42Z" level=debug msg="elasticsearch/artifacts -> elasticsearch_artifacts" time="2025-02-17T20:26:42Z" level=debug msg="elasticsearch/es-multiinstance -> elasticsearch_es_multiinstance" time="2025-02-17T20:26:42Z" level=debug msg="elasticsearch/es-simple-prod -> elasticsearch_es_simple_prod" time="2025-02-17T20:26:42Z" level=debug msg="elasticsearch/es-rollover-autoprov -> elasticsearch_es_rollover_autoprov" time="2025-02-17T20:26:42Z" level=debug msg="elasticsearch/es-increasing-replicas -> elasticsearch_es_increasing_replicas" time="2025-02-17T20:26:42Z" level=debug msg="elasticsearch/es-index-cleaner-autoprov -> elasticsearch_es_index_cleaner_autoprov" time="2025-02-17T20:26:42Z" level=debug msg="elasticsearch/es-from-aio-to-production -> elasticsearch_es_from_aio_to_production" +-----------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------+--------+ | elasticsearch_artifacts | passed | | elasticsearch_es_multiinstance | passed | | elasticsearch_es_simple_prod | passed | | elasticsearch_es_rollover_autoprov | passed | | elasticsearch_es_increasing_replicas | passed | | elasticsearch_es_index_cleaner_autoprov | passed | | elasticsearch_es_from_aio_to_production | passed | +-----------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh examples false true + '[' 3 -ne 3 ']' + test_suite_name=examples + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/examples.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-examples make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ VERTX_IMG=jaegertracing/vertx-create-span:operator-e2e-tests \ ./tests/e2e/examples/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-14-222249 True False 23m Cluster version is 4.18.0-0.nightly-2025-02-14-222249' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-14-222249 True False 23m Cluster version is 4.18.0-0.nightly-2025-02-14-222249' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/examples/render.sh ++ export SUITE_DIR=./tests/e2e/examples ++ SUITE_DIR=./tests/e2e/examples ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/examples ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test examples-agent-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-agent-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-with-priority-class\e[0m' Rendering files for test examples-agent-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + mkdir -p examples-agent-with-priority-class + cd examples-agent-with-priority-class + example_name=agent-with-priority-class + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + '[' true '!=' true ']' + render_install_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + install_file=./02-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/agent-with-priority-class.yaml -o ./02-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./02-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./02-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./02-install.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + JAEGER_NAME=agent-as-daemonset + local jaeger_strategy ++ get_jaeger_strategy ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./02-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./02-install.yaml ++ strategy=DaemonSet ++ '[' DaemonSet = null ']' ++ echo DaemonSet ++ return 0 + jaeger_strategy=DaemonSet + '[' DaemonSet = DaemonSet ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./02-assert.yaml + render_smoke_test_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + jaeger_name=agent-as-daemonset + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test agent-as-daemonset true 02 + '[' 3 -ne 3 ']' + jaeger=agent-as-daemonset + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-all-in-one-with-options + '[' 1 -ne 1 ']' + test_name=examples-all-in-one-with-options + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-all-in-one-with-options' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-all-in-one-with-options\e[0m' Rendering files for test examples-all-in-one-with-options + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-with-priority-class + '[' examples-agent-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-all-in-one-with-options + cd examples-all-in-one-with-options + example_name=all-in-one-with-options + render_install_example all-in-one-with-options 00 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/all-in-one-with-options.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + JAEGER_NAME=my-jaeger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.metadata.name="my-jaeger"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i 'del(.spec.allInOne.image)' ./00-install.yaml + render_smoke_test_example all-in-one-with-options 01 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + jaeger_name=my-jaeger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + sed -i s~my-jaeger-query:443~my-jaeger-query:443/jaeger~gi ./01-smoke-test.yaml + start_test examples-business-application-injected-sidecar + '[' 1 -ne 1 ']' + test_name=examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-business-application-injected-sidecar' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-business-application-injected-sidecar\e[0m' Rendering files for test examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-all-in-one-with-options + '[' examples-all-in-one-with-options '!=' _build ']' + cd .. + mkdir -p examples-business-application-injected-sidecar + cd examples-business-application-injected-sidecar + example_name=simplest + cp /tmp/jaeger-tests/examples/business-application-injected-sidecar.yaml ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].image=strenv(VERTX_IMG)' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.path="/"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.port=8080' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.initialDelaySeconds=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.failureThreshold=3' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.periodSeconds=10' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.successThreshold=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.timeoutSeconds=1' ./00-install.yaml + render_install_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example simplest 02 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 02 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-collector-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-collector-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-collector-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-collector-with-priority-class\e[0m' Rendering files for test examples-collector-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-business-application-injected-sidecar + '[' examples-business-application-injected-sidecar '!=' _build ']' + cd .. + mkdir -p examples-collector-with-priority-class + cd examples-collector-with-priority-class + example_name=collector-with-priority-class + render_install_example collector-with-priority-class 00 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/collector-with-priority-class.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + JAEGER_NAME=collector-with-high-priority + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example collector-with-priority-class 01 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + jaeger_name=collector-with-high-priority + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test collector-with-high-priority true 01 + '[' 3 -ne 3 ']' + jaeger=collector-with-high-priority + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + export JAEGER_NAME=collector-with-high-priority + JAEGER_NAME=collector-with-high-priority + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-service-types + '[' 1 -ne 1 ']' + test_name=examples-service-types + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-service-types' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-service-types\e[0m' Rendering files for test examples-service-types + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-collector-with-priority-class + '[' examples-collector-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-service-types + cd examples-service-types + example_name=service-types + render_install_example service-types 00 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/service-types.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + JAEGER_NAME=service-types + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example service-types 01 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/service-types.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/service-types.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/service-types.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/service-types.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + jaeger_name=service-types + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test service-types true 01 + '[' 3 -ne 3 ']' + jaeger=service-types + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + export JAEGER_NAME=service-types + JAEGER_NAME=service-types + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod + '[' 1 -ne 1 ']' + test_name=examples-simple-prod + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod\e[0m' Rendering files for test examples-simple-prod + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-service-types + '[' examples-service-types '!=' _build ']' + cd .. + mkdir -p examples-simple-prod + cd examples-simple-prod + example_name=simple-prod + render_install_example simple-prod 01 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod 02 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod-with-volumes + '[' 1 -ne 1 ']' + test_name=examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod-with-volumes' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod-with-volumes\e[0m' Rendering files for test examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod + '[' examples-simple-prod '!=' _build ']' + cd .. + mkdir -p examples-simple-prod-with-volumes + cd examples-simple-prod-with-volumes + example_name=simple-prod-with-volumes + render_install_example simple-prod-with-volumes 01 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod-with-volumes 02 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + /tmp/jaeger-tests/bin/gomplate -f ./03-check-volume.yaml.template -o 03-check-volume.yaml + start_test examples-simplest + '[' 1 -ne 1 ']' + test_name=examples-simplest + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simplest' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simplest\e[0m' Rendering files for test examples-simplest + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod-with-volumes + '[' examples-simple-prod-with-volumes '!=' _build ']' + cd .. + mkdir -p examples-simplest + cd examples-simplest + example_name=simplest + render_install_example simplest 00 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 01 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger + '[' 1 -ne 1 ']' + test_name=examples-with-badger + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger\e[0m' Rendering files for test examples-with-badger + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simplest + '[' examples-simplest '!=' _build ']' + cd .. + mkdir -p examples-with-badger + cd examples-with-badger + example_name=with-badger + render_install_example with-badger 00 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + JAEGER_NAME=with-badger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger 01 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + jaeger_name=with-badger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + export JAEGER_NAME=with-badger + JAEGER_NAME=with-badger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger-and-volume + '[' 1 -ne 1 ']' + test_name=examples-with-badger-and-volume + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger-and-volume' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger-and-volume\e[0m' Rendering files for test examples-with-badger-and-volume + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger + '[' examples-with-badger '!=' _build ']' + cd .. + mkdir -p examples-with-badger-and-volume + cd examples-with-badger-and-volume + example_name=with-badger-and-volume + render_install_example with-badger-and-volume 00 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger-and-volume.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + JAEGER_NAME=with-badger-and-volume + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger-and-volume 01 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + jaeger_name=with-badger-and-volume + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger-and-volume true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger-and-volume + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + export JAEGER_NAME=with-badger-and-volume + JAEGER_NAME=with-badger-and-volume + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-cassandra + '[' 1 -ne 1 ']' + test_name=examples-with-cassandra + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-cassandra' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-cassandra\e[0m' Rendering files for test examples-with-cassandra + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger-and-volume + '[' examples-with-badger-and-volume '!=' _build ']' + cd .. + mkdir -p examples-with-cassandra + cd examples-with-cassandra + example_name=with-cassandra + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-cassandra 01 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-cassandra.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + JAEGER_NAME=with-cassandra + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-cassandra 02 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-cassandra.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-cassandra.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + jaeger_name=with-cassandra + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-cassandra true 02 + '[' 3 -ne 3 ']' + jaeger=with-cassandra + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + export JAEGER_NAME=with-cassandra + JAEGER_NAME=with-cassandra + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-sampling + '[' 1 -ne 1 ']' + test_name=examples-with-sampling + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-sampling' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-sampling\e[0m' Rendering files for test examples-with-sampling + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-cassandra + '[' examples-with-cassandra '!=' _build ']' + cd .. + mkdir -p examples-with-sampling + cd examples-with-sampling + export example_name=with-sampling + example_name=with-sampling + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-sampling 01 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-sampling.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + JAEGER_NAME=with-sampling + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-sampling 02 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-sampling.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-sampling.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + jaeger_name=with-sampling + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-sampling true 02 + '[' 3 -ne 3 ']' + jaeger=with-sampling + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + export JAEGER_NAME=with-sampling + JAEGER_NAME=with-sampling + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-agent-as-daemonset + '[' 1 -ne 1 ']' + test_name=examples-agent-as-daemonset + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-as-daemonset' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-as-daemonset\e[0m' Rendering files for test examples-agent-as-daemonset + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-sampling + '[' examples-with-sampling '!=' _build ']' + cd .. + mkdir -p examples-agent-as-daemonset + cd examples-agent-as-daemonset + '[' true = true ']' + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/agent-as-daemonset.yaml -o 02-install.yaml + '[' true = true ']' + start_test examples-openshift-with-htpasswd + '[' 1 -ne 1 ']' + test_name=examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-openshift-with-htpasswd' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-openshift-with-htpasswd\e[0m' Rendering files for test examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-as-daemonset + '[' examples-agent-as-daemonset '!=' _build ']' + cd .. + mkdir -p examples-openshift-with-htpasswd + cd examples-openshift-with-htpasswd + export JAEGER_NAME=with-htpasswd + JAEGER_NAME=with-htpasswd + export JAEGER_USERNAME=awesomeuser + JAEGER_USERNAME=awesomeuser + export JAEGER_PASSWORD=awesomepassword + JAEGER_PASSWORD=awesomepassword + export 'JAEGER_USER_PASSWORD_HASH=awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' + JAEGER_USER_PASSWORD_HASH='awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ echo 'awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ base64 + SECRET=YXdlc29tZXVzZXI6e1NIQX11VWRxUFZVeXFOQm1FUlUwUXhqM0tGYVpuanc9Cg== + /tmp/jaeger-tests/bin/gomplate -f ./00-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/with-htpasswd.yaml -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + export 'GET_URL_COMMAND=kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + export 'URL=https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + INSECURE=true + JAEGER_USERNAME= + JAEGER_PASSWORD= + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-unsecured.yaml + JAEGER_USERNAME=wronguser + JAEGER_PASSWORD=wrongpassword + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-check-unauthorized.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./04-check-authorized.yaml + skip_test examples-agent-as-daemonset 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-agent-as-daemonset + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-openshift-with-htpasswd + '[' examples-openshift-with-htpasswd '!=' _build ']' + cd .. + rm -rf examples-agent-as-daemonset + warning 'examples-agent-as-daemonset: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-agent-as-daemonset: This test is flaky in Prow CI\e[0m' WAR: examples-agent-as-daemonset: This test is flaky in Prow CI + skip_test examples-with-badger-and-volume 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-with-badger-and-volume + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-with-badger-and-volume + warning 'examples-with-badger-and-volume: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-with-badger-and-volume: This test is flaky in Prow CI\e[0m' WAR: examples-with-badger-and-volume: This test is flaky in Prow CI + skip_test examples-collector-with-priority-class 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-collector-with-priority-class + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-collector-with-priority-class + warning 'examples-collector-with-priority-class: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-collector-with-priority-class: This test is flaky in Prow CI\e[0m' WAR: examples-collector-with-priority-class: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running examples E2E tests' Running examples E2E tests + cd tests/e2e/examples/_build + set +e + KUBECONFIG=/tmp/kubeconfig-998185858 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 12 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/examples-agent-with-priority-class === PAUSE kuttl/harness/examples-agent-with-priority-class === RUN kuttl/harness/examples-all-in-one-with-options === PAUSE kuttl/harness/examples-all-in-one-with-options === RUN kuttl/harness/examples-business-application-injected-sidecar === PAUSE kuttl/harness/examples-business-application-injected-sidecar === RUN kuttl/harness/examples-openshift-with-htpasswd === PAUSE kuttl/harness/examples-openshift-with-htpasswd === RUN kuttl/harness/examples-service-types === PAUSE kuttl/harness/examples-service-types === RUN kuttl/harness/examples-simple-prod === PAUSE kuttl/harness/examples-simple-prod === RUN kuttl/harness/examples-simple-prod-with-volumes === PAUSE kuttl/harness/examples-simple-prod-with-volumes === RUN kuttl/harness/examples-simplest === PAUSE kuttl/harness/examples-simplest === RUN kuttl/harness/examples-with-badger === PAUSE kuttl/harness/examples-with-badger === RUN kuttl/harness/examples-with-cassandra === PAUSE kuttl/harness/examples-with-cassandra === RUN kuttl/harness/examples-with-sampling === PAUSE kuttl/harness/examples-with-sampling === CONT kuttl/harness/artifacts logger.go:42: 20:27:13 | artifacts | Creating namespace: kuttl-test-touching-cow logger.go:42: 20:27:13 | artifacts | artifacts events from ns kuttl-test-touching-cow: logger.go:42: 20:27:13 | artifacts | Deleting namespace: kuttl-test-touching-cow === CONT kuttl/harness/examples-simple-prod logger.go:42: 20:27:19 | examples-simple-prod | Creating namespace: kuttl-test-regular-raven logger.go:42: 20:27:19 | examples-simple-prod/1-install | starting test step 1-install logger.go:42: 20:27:19 | examples-simple-prod/1-install | Jaeger:kuttl-test-regular-raven/simple-prod created logger.go:42: 20:27:56 | examples-simple-prod/1-install | test step completed 1-install logger.go:42: 20:27:56 | examples-simple-prod/2-smoke-test | starting test step 2-smoke-test logger.go:42: 20:27:56 | examples-simple-prod/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 20:27:57 | examples-simple-prod/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:28:05 | examples-simple-prod/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 20:28:06 | examples-simple-prod/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 20:28:06 | examples-simple-prod/2-smoke-test | job.batch/report-span created logger.go:42: 20:28:06 | examples-simple-prod/2-smoke-test | job.batch/check-span created logger.go:42: 20:28:18 | examples-simple-prod/2-smoke-test | test step completed 2-smoke-test logger.go:42: 20:28:18 | examples-simple-prod | examples-simple-prod events from ns kuttl-test-regular-raven: logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:25 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestregularravensimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestregularravensimpleprod-1-64654c9c44 to 1 deployment-controller logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:26 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestregularravensimpleprod-1-64654c9c44 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestregularravensimpleprod-1-64654cjrsc9 replicaset-controller logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestregularravensimpleprod-1-64654cjrsc9 Binding Scheduled Successfully assigned kuttl-test-regular-raven/elasticsearch-cdm-kuttltestregularravensimpleprod-1-64654cjrsc9 to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestregularravensimpleprod-1-64654cjrsc9 AddedInterface Add eth0 [10.129.2.32/23] from ovn-kubernetes multus logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestregularravensimpleprod-1-64654cjrsc9.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:5117d0aee260f74d6ef77c7c71ed073b1ad45b456f58894539b5697bec878310" already present on machine kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestregularravensimpleprod-1-64654cjrsc9.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestregularravensimpleprod-1-64654cjrsc9.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestregularravensimpleprod-1-64654cjrsc9.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d63d612cdca72f64c45960a9b30bcfd0c746daabc17407df92c62694d5bc4bb6" already present on machine kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestregularravensimpleprod-1-64654cjrsc9.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestregularravensimpleprod-1-64654cjrsc9.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:36 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestregularravensimpleprod-1-64654cjrsc9.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:41 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestregularravensimpleprod-1-64654cjrsc9.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:52 +0000 UTC Normal Pod simple-prod-collector-5d479b4796-t464r Binding Scheduled Successfully assigned kuttl-test-regular-raven/simple-prod-collector-5d479b4796-t464r to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:52 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5d479b4796 SuccessfulCreate Created pod: simple-prod-collector-5d479b4796-t464r replicaset-controller logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:52 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-5d479b4796 to 1 deployment-controller logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:52 +0000 UTC Normal Pod simple-prod-query-85c8569495-kcjjf Binding Scheduled Successfully assigned kuttl-test-regular-raven/simple-prod-query-85c8569495-kcjjf to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:52 +0000 UTC Normal ReplicaSet.apps simple-prod-query-85c8569495 SuccessfulCreate Created pod: simple-prod-query-85c8569495-kcjjf replicaset-controller logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:52 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-85c8569495 to 1 deployment-controller logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:53 +0000 UTC Normal Pod simple-prod-collector-5d479b4796-t464r AddedInterface Add eth0 [10.131.0.48/23] from ovn-kubernetes multus logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:53 +0000 UTC Normal Pod simple-prod-collector-5d479b4796-t464r.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:53 +0000 UTC Normal Pod simple-prod-collector-5d479b4796-t464r.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:53 +0000 UTC Normal Pod simple-prod-collector-5d479b4796-t464r.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:53 +0000 UTC Normal Pod simple-prod-query-85c8569495-kcjjf AddedInterface Add eth0 [10.128.2.35/23] from ovn-kubernetes multus logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:53 +0000 UTC Normal Pod simple-prod-query-85c8569495-kcjjf.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:53 +0000 UTC Normal Pod simple-prod-query-85c8569495-kcjjf.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:53 +0000 UTC Normal Pod simple-prod-query-85c8569495-kcjjf.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:53 +0000 UTC Normal Pod simple-prod-query-85c8569495-kcjjf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:53 +0000 UTC Normal Pod simple-prod-query-85c8569495-kcjjf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:53 +0000 UTC Normal Pod simple-prod-query-85c8569495-kcjjf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:53 +0000 UTC Normal Pod simple-prod-query-85c8569495-kcjjf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:53 +0000 UTC Normal Pod simple-prod-query-85c8569495-kcjjf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:27:53 +0000 UTC Normal Pod simple-prod-query-85c8569495-kcjjf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:02 +0000 UTC Normal Pod simple-prod-query-85c8569495-kcjjf.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:02 +0000 UTC Normal Pod simple-prod-query-85c8569495-kcjjf.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:02 +0000 UTC Normal Pod simple-prod-query-85c8569495-kcjjf.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:02 +0000 UTC Normal ReplicaSet.apps simple-prod-query-85c8569495 SuccessfulDelete Deleted pod: simple-prod-query-85c8569495-kcjjf replicaset-controller logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:02 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-85c8569495 to 0 from 1 deployment-controller logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:03 +0000 UTC Normal Pod simple-prod-query-58cbf8f47b-nmltx Binding Scheduled Successfully assigned kuttl-test-regular-raven/simple-prod-query-58cbf8f47b-nmltx to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:03 +0000 UTC Normal ReplicaSet.apps simple-prod-query-58cbf8f47b SuccessfulCreate Created pod: simple-prod-query-58cbf8f47b-nmltx replicaset-controller logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:03 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-58cbf8f47b to 1 deployment-controller logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:04 +0000 UTC Normal Pod simple-prod-query-58cbf8f47b-nmltx AddedInterface Add eth0 [10.128.2.36/23] from ovn-kubernetes multus logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:04 +0000 UTC Normal Pod simple-prod-query-58cbf8f47b-nmltx.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:04 +0000 UTC Normal Pod simple-prod-query-58cbf8f47b-nmltx.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:04 +0000 UTC Normal Pod simple-prod-query-58cbf8f47b-nmltx.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:04 +0000 UTC Normal Pod simple-prod-query-58cbf8f47b-nmltx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:04 +0000 UTC Normal Pod simple-prod-query-58cbf8f47b-nmltx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:04 +0000 UTC Normal Pod simple-prod-query-58cbf8f47b-nmltx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:04 +0000 UTC Normal Pod simple-prod-query-58cbf8f47b-nmltx.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:04 +0000 UTC Normal Pod simple-prod-query-58cbf8f47b-nmltx.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:04 +0000 UTC Normal Pod simple-prod-query-58cbf8f47b-nmltx.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:06 +0000 UTC Normal Pod check-span-psp58 Binding Scheduled Successfully assigned kuttl-test-regular-raven/check-span-psp58 to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:06 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-psp58 job-controller logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:06 +0000 UTC Normal Pod report-span-hk8g7 Binding Scheduled Successfully assigned kuttl-test-regular-raven/report-span-hk8g7 to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:06 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-hk8g7 job-controller logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:07 +0000 UTC Normal Pod check-span-psp58 AddedInterface Add eth0 [10.131.0.50/23] from ovn-kubernetes multus logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:07 +0000 UTC Normal Pod check-span-psp58.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:07 +0000 UTC Normal Pod check-span-psp58.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 593ms (593ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:07 +0000 UTC Normal Pod check-span-psp58.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:07 +0000 UTC Normal Pod check-span-psp58.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:07 +0000 UTC Normal Pod report-span-hk8g7 AddedInterface Add eth0 [10.131.0.49/23] from ovn-kubernetes multus logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:07 +0000 UTC Normal Pod report-span-hk8g7.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:08 +0000 UTC Normal Pod report-span-hk8g7.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 953ms (953ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:08 +0000 UTC Normal Pod report-span-hk8g7.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:08 +0000 UTC Normal Pod report-span-hk8g7.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:28:18 | examples-simple-prod | 2025-02-17 20:28:18 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 20:28:18 | examples-simple-prod | Deleting namespace: kuttl-test-regular-raven === CONT kuttl/harness/examples-with-sampling logger.go:42: 20:28:31 | examples-with-sampling | Creating namespace: kuttl-test-calm-anemone logger.go:42: 20:28:32 | examples-with-sampling/0-install | starting test step 0-install logger.go:42: 20:28:32 | examples-with-sampling/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 20:28:32 | examples-with-sampling/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 20:28:32 | examples-with-sampling/0-install | >>>> Creating namespace kuttl-test-calm-anemone logger.go:42: 20:28:32 | examples-with-sampling/0-install | kubectl create namespace kuttl-test-calm-anemone 2>&1 | grep -v "already exists" || true logger.go:42: 20:28:32 | examples-with-sampling/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-calm-anemone 2>&1 | grep -v "already exists" || true logger.go:42: 20:28:32 | examples-with-sampling/0-install | service/cassandra created logger.go:42: 20:28:32 | examples-with-sampling/0-install | statefulset.apps/cassandra created logger.go:42: 20:28:32 | examples-with-sampling/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 20:28:44 | examples-with-sampling/0-install | test step completed 0-install logger.go:42: 20:28:44 | examples-with-sampling/1-install | starting test step 1-install logger.go:42: 20:28:44 | examples-with-sampling/1-install | Jaeger:kuttl-test-calm-anemone/with-sampling created logger.go:42: 20:28:51 | examples-with-sampling/1-install | test step completed 1-install logger.go:42: 20:28:51 | examples-with-sampling/2-smoke-test | starting test step 2-smoke-test logger.go:42: 20:28:51 | examples-with-sampling/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-sampling /dev/null] logger.go:42: 20:28:52 | examples-with-sampling/2-smoke-test | Warning: resource jaegers/with-sampling is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:28:59 | examples-with-sampling/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 20:29:00 | examples-with-sampling/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 20:29:00 | examples-with-sampling/2-smoke-test | job.batch/report-span created logger.go:42: 20:29:00 | examples-with-sampling/2-smoke-test | job.batch/check-span created logger.go:42: 20:29:12 | examples-with-sampling/2-smoke-test | test step completed 2-smoke-test logger.go:42: 20:29:12 | examples-with-sampling/3- | starting test step 3- logger.go:42: 20:29:12 | examples-with-sampling/3- | test step completed 3- logger.go:42: 20:29:12 | examples-with-sampling | examples-with-sampling events from ns kuttl-test-calm-anemone: logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:32 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-calm-anemone/cassandra-0 to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:32 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:33 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.129.2.33/23] from ovn-kubernetes multus logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:33 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:37 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 3.854s (3.854s including waiting). Image size: 309834747 bytes. kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:37 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:37 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:38 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-calm-anemone/cassandra-1 to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:38 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.131.0.51/23] from ovn-kubernetes multus logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:38 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:38 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:43 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 4.939s (4.939s including waiting). Image size: 309834747 bytes. kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:43 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:43 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:49 +0000 UTC Normal Pod with-sampling-67b749c5bb-f5mht Binding Scheduled Successfully assigned kuttl-test-calm-anemone/with-sampling-67b749c5bb-f5mht to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:49 +0000 UTC Normal Pod with-sampling-67b749c5bb-f5mht AddedInterface Add eth0 [10.128.2.37/23] from ovn-kubernetes multus logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:49 +0000 UTC Normal Pod with-sampling-67b749c5bb-f5mht.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:49 +0000 UTC Normal Pod with-sampling-67b749c5bb-f5mht.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:49 +0000 UTC Normal Pod with-sampling-67b749c5bb-f5mht.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:49 +0000 UTC Normal Pod with-sampling-67b749c5bb-f5mht.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:49 +0000 UTC Normal Pod with-sampling-67b749c5bb-f5mht.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:49 +0000 UTC Normal Pod with-sampling-67b749c5bb-f5mht.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:49 +0000 UTC Normal ReplicaSet.apps with-sampling-67b749c5bb SuccessfulCreate Created pod: with-sampling-67b749c5bb-f5mht replicaset-controller logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:49 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-67b749c5bb to 1 deployment-controller logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:56 +0000 UTC Normal Pod with-sampling-67b749c5bb-f5mht.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:56 +0000 UTC Normal Pod with-sampling-67b749c5bb-f5mht.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:56 +0000 UTC Normal ReplicaSet.apps with-sampling-67b749c5bb SuccessfulDelete Deleted pod: with-sampling-67b749c5bb-f5mht replicaset-controller logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:56 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled down replica set with-sampling-67b749c5bb to 0 from 1 deployment-controller logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:57 +0000 UTC Normal ReplicaSet.apps with-sampling-5b88d957cb SuccessfulCreate Created pod: with-sampling-5b88d957cb-6svn9 replicaset-controller logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:57 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-5b88d957cb to 1 deployment-controller logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:58 +0000 UTC Normal Pod with-sampling-5b88d957cb-6svn9 Binding Scheduled Successfully assigned kuttl-test-calm-anemone/with-sampling-5b88d957cb-6svn9 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:58 +0000 UTC Normal Pod with-sampling-5b88d957cb-6svn9 AddedInterface Add eth0 [10.128.2.38/23] from ovn-kubernetes multus logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:58 +0000 UTC Normal Pod with-sampling-5b88d957cb-6svn9.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:58 +0000 UTC Normal Pod with-sampling-5b88d957cb-6svn9.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:58 +0000 UTC Normal Pod with-sampling-5b88d957cb-6svn9.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:58 +0000 UTC Normal Pod with-sampling-5b88d957cb-6svn9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:58 +0000 UTC Normal Pod with-sampling-5b88d957cb-6svn9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:28:58 +0000 UTC Normal Pod with-sampling-5b88d957cb-6svn9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:29:00 +0000 UTC Normal Pod check-span-pxnzl Binding Scheduled Successfully assigned kuttl-test-calm-anemone/check-span-pxnzl to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:29:00 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-pxnzl job-controller logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:29:00 +0000 UTC Normal Pod report-span-8ggwp Binding Scheduled Successfully assigned kuttl-test-calm-anemone/report-span-8ggwp to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:29:00 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-8ggwp job-controller logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:29:01 +0000 UTC Normal Pod check-span-pxnzl AddedInterface Add eth0 [10.131.0.52/23] from ovn-kubernetes multus logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:29:01 +0000 UTC Normal Pod check-span-pxnzl.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:29:01 +0000 UTC Normal Pod check-span-pxnzl.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 688ms (689ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:29:01 +0000 UTC Normal Pod report-span-8ggwp AddedInterface Add eth0 [10.129.2.34/23] from ovn-kubernetes multus logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:29:01 +0000 UTC Normal Pod report-span-8ggwp.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:29:02 +0000 UTC Normal Pod check-span-pxnzl.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:29:02 +0000 UTC Normal Pod check-span-pxnzl.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:29:02 +0000 UTC Normal Pod report-span-8ggwp.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.198s (1.198s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:29:02 +0000 UTC Normal Pod report-span-8ggwp.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:29:02 +0000 UTC Normal Pod report-span-8ggwp.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 20:29:12 | examples-with-sampling | 2025-02-17 20:29:12 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 20:29:12 | examples-with-sampling | Deleting namespace: kuttl-test-calm-anemone === CONT kuttl/harness/examples-with-cassandra logger.go:42: 20:29:31 | examples-with-cassandra | Creating namespace: kuttl-test-smiling-wolf logger.go:42: 20:29:31 | examples-with-cassandra/0-install | starting test step 0-install logger.go:42: 20:29:31 | examples-with-cassandra/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 20:29:31 | examples-with-cassandra/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 20:29:31 | examples-with-cassandra/0-install | >>>> Creating namespace kuttl-test-smiling-wolf logger.go:42: 20:29:31 | examples-with-cassandra/0-install | kubectl create namespace kuttl-test-smiling-wolf 2>&1 | grep -v "already exists" || true logger.go:42: 20:29:31 | examples-with-cassandra/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-smiling-wolf 2>&1 | grep -v "already exists" || true logger.go:42: 20:29:32 | examples-with-cassandra/0-install | service/cassandra created logger.go:42: 20:29:32 | examples-with-cassandra/0-install | statefulset.apps/cassandra created logger.go:42: 20:29:32 | examples-with-cassandra/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 20:29:34 | examples-with-cassandra/0-install | test step completed 0-install logger.go:42: 20:29:34 | examples-with-cassandra/1-install | starting test step 1-install logger.go:42: 20:29:34 | examples-with-cassandra/1-install | Jaeger:kuttl-test-smiling-wolf/with-cassandra created logger.go:42: 20:29:53 | examples-with-cassandra/1-install | test step completed 1-install logger.go:42: 20:29:53 | examples-with-cassandra/2-smoke-test | starting test step 2-smoke-test logger.go:42: 20:29:53 | examples-with-cassandra/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-cassandra /dev/null] logger.go:42: 20:29:55 | examples-with-cassandra/2-smoke-test | Warning: resource jaegers/with-cassandra is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:30:01 | examples-with-cassandra/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 20:30:02 | examples-with-cassandra/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 20:30:02 | examples-with-cassandra/2-smoke-test | job.batch/report-span created logger.go:42: 20:30:02 | examples-with-cassandra/2-smoke-test | job.batch/check-span created logger.go:42: 20:30:14 | examples-with-cassandra/2-smoke-test | test step completed 2-smoke-test logger.go:42: 20:30:14 | examples-with-cassandra | examples-with-cassandra events from ns kuttl-test-smiling-wolf: logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:32 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-smiling-wolf/cassandra-0 to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:32 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:33 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.129.2.35/23] from ovn-kubernetes multus logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:33 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:33 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:33 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:33 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-smiling-wolf/cassandra-1 to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:33 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.131.0.53/23] from ovn-kubernetes multus logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:33 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:33 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:34 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:34 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:38 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-gwpd5 Binding Scheduled Successfully assigned kuttl-test-smiling-wolf/with-cassandra-cassandra-schema-job-gwpd5 to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:38 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-gwpd5 AddedInterface Add eth0 [10.129.2.36/23] from ovn-kubernetes multus logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:38 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-gwpd5.spec.containers{with-cassandra-cassandra-schema-job} Pulling Pulling image "jaegertracing/jaeger-cassandra-schema:1.62.0" kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:38 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job SuccessfulCreate Created pod: with-cassandra-cassandra-schema-job-gwpd5 job-controller logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:43 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-gwpd5.spec.containers{with-cassandra-cassandra-schema-job} Pulled Successfully pulled image "jaegertracing/jaeger-cassandra-schema:1.62.0" in 4.483s (4.483s including waiting). Image size: 405965365 bytes. kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:43 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-gwpd5.spec.containers{with-cassandra-cassandra-schema-job} Created Created container with-cassandra-cassandra-schema-job kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:43 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-gwpd5.spec.containers{with-cassandra-cassandra-schema-job} Started Started container with-cassandra-cassandra-schema-job kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:50 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job Completed Job completed job-controller logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:51 +0000 UTC Normal Pod with-cassandra-6694b66d5c-clkk6 Binding Scheduled Successfully assigned kuttl-test-smiling-wolf/with-cassandra-6694b66d5c-clkk6 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:51 +0000 UTC Normal ReplicaSet.apps with-cassandra-6694b66d5c SuccessfulCreate Created pod: with-cassandra-6694b66d5c-clkk6 replicaset-controller logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:51 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-6694b66d5c to 1 deployment-controller logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:52 +0000 UTC Normal Pod with-cassandra-6694b66d5c-clkk6 AddedInterface Add eth0 [10.128.2.39/23] from ovn-kubernetes multus logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:52 +0000 UTC Normal Pod with-cassandra-6694b66d5c-clkk6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:52 +0000 UTC Normal Pod with-cassandra-6694b66d5c-clkk6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:52 +0000 UTC Normal Pod with-cassandra-6694b66d5c-clkk6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:52 +0000 UTC Normal Pod with-cassandra-6694b66d5c-clkk6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:52 +0000 UTC Normal Pod with-cassandra-6694b66d5c-clkk6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:52 +0000 UTC Normal Pod with-cassandra-6694b66d5c-clkk6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:58 +0000 UTC Normal Pod with-cassandra-6694b66d5c-clkk6.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:58 +0000 UTC Normal Pod with-cassandra-6694b66d5c-clkk6.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:58 +0000 UTC Normal ReplicaSet.apps with-cassandra-6694b66d5c SuccessfulDelete Deleted pod: with-cassandra-6694b66d5c-clkk6 replicaset-controller logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:58 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled down replica set with-cassandra-6694b66d5c to 0 from 1 deployment-controller logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:59 +0000 UTC Normal Pod with-cassandra-b7b69785d-tpl6x Binding Scheduled Successfully assigned kuttl-test-smiling-wolf/with-cassandra-b7b69785d-tpl6x to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:59 +0000 UTC Normal Pod with-cassandra-b7b69785d-tpl6x AddedInterface Add eth0 [10.128.2.40/23] from ovn-kubernetes multus logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:59 +0000 UTC Normal Pod with-cassandra-b7b69785d-tpl6x.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:59 +0000 UTC Normal Pod with-cassandra-b7b69785d-tpl6x.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:59 +0000 UTC Normal Pod with-cassandra-b7b69785d-tpl6x.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:59 +0000 UTC Normal Pod with-cassandra-b7b69785d-tpl6x.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:59 +0000 UTC Normal Pod with-cassandra-b7b69785d-tpl6x.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:59 +0000 UTC Normal Pod with-cassandra-b7b69785d-tpl6x.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:59 +0000 UTC Normal ReplicaSet.apps with-cassandra-b7b69785d SuccessfulCreate Created pod: with-cassandra-b7b69785d-tpl6x replicaset-controller logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:29:59 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-b7b69785d to 1 deployment-controller logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:30:02 +0000 UTC Normal Pod check-span-7k4vz Binding Scheduled Successfully assigned kuttl-test-smiling-wolf/check-span-7k4vz to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:30:02 +0000 UTC Normal Pod check-span-7k4vz AddedInterface Add eth0 [10.129.2.39/23] from ovn-kubernetes multus logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:30:02 +0000 UTC Normal Pod check-span-7k4vz.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:30:02 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-7k4vz job-controller logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:30:02 +0000 UTC Normal Pod report-span-bdmxn Binding Scheduled Successfully assigned kuttl-test-smiling-wolf/report-span-bdmxn to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:30:02 +0000 UTC Normal Pod report-span-bdmxn AddedInterface Add eth0 [10.129.2.38/23] from ovn-kubernetes multus logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:30:02 +0000 UTC Normal Pod report-span-bdmxn.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:30:02 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-bdmxn job-controller logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:30:03 +0000 UTC Normal Pod check-span-7k4vz.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 841ms (841ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:30:03 +0000 UTC Normal Pod check-span-7k4vz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:30:03 +0000 UTC Normal Pod check-span-7k4vz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:30:04 +0000 UTC Normal Pod report-span-bdmxn.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.436s (1.436s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:30:04 +0000 UTC Normal Pod report-span-bdmxn.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:30:04 +0000 UTC Normal Pod report-span-bdmxn.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 20:30:14 | examples-with-cassandra | 2025-02-17 20:30:14 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 20:30:14 | examples-with-cassandra | Deleting namespace: kuttl-test-smiling-wolf === CONT kuttl/harness/examples-with-badger logger.go:42: 20:30:27 | examples-with-badger | Creating namespace: kuttl-test-more-goldfish logger.go:42: 20:30:27 | examples-with-badger/0-install | starting test step 0-install logger.go:42: 20:30:28 | examples-with-badger/0-install | Jaeger:kuttl-test-more-goldfish/with-badger created logger.go:42: 20:30:34 | examples-with-badger/0-install | test step completed 0-install logger.go:42: 20:30:34 | examples-with-badger/1-smoke-test | starting test step 1-smoke-test logger.go:42: 20:30:34 | examples-with-badger/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger /dev/null] logger.go:42: 20:30:35 | examples-with-badger/1-smoke-test | Warning: resource jaegers/with-badger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:30:42 | examples-with-badger/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 20:30:42 | examples-with-badger/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 20:30:42 | examples-with-badger/1-smoke-test | job.batch/report-span created logger.go:42: 20:30:42 | examples-with-badger/1-smoke-test | job.batch/check-span created logger.go:42: 20:30:56 | examples-with-badger/1-smoke-test | test step completed 1-smoke-test logger.go:42: 20:30:56 | examples-with-badger | examples-with-badger events from ns kuttl-test-more-goldfish: logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:31 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-7c5675565c to 1 deployment-controller logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:32 +0000 UTC Normal Pod with-badger-7c5675565c-5j5t8 Binding Scheduled Successfully assigned kuttl-test-more-goldfish/with-badger-7c5675565c-5j5t8 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:32 +0000 UTC Normal Pod with-badger-7c5675565c-5j5t8 AddedInterface Add eth0 [10.128.2.41/23] from ovn-kubernetes multus logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:32 +0000 UTC Normal Pod with-badger-7c5675565c-5j5t8.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:32 +0000 UTC Normal Pod with-badger-7c5675565c-5j5t8.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:32 +0000 UTC Normal Pod with-badger-7c5675565c-5j5t8.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:32 +0000 UTC Normal Pod with-badger-7c5675565c-5j5t8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:32 +0000 UTC Normal Pod with-badger-7c5675565c-5j5t8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:32 +0000 UTC Normal Pod with-badger-7c5675565c-5j5t8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:32 +0000 UTC Normal ReplicaSet.apps with-badger-7c5675565c SuccessfulCreate Created pod: with-badger-7c5675565c-5j5t8 replicaset-controller logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:39 +0000 UTC Normal Pod with-badger-7c5675565c-5j5t8.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:39 +0000 UTC Normal Pod with-badger-7c5675565c-5j5t8.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:39 +0000 UTC Normal ReplicaSet.apps with-badger-7c5675565c SuccessfulDelete Deleted pod: with-badger-7c5675565c-5j5t8 replicaset-controller logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:39 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled down replica set with-badger-7c5675565c to 0 from 1 deployment-controller logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:41 +0000 UTC Normal Pod with-badger-7b9bfb4788-96v7d Binding Scheduled Successfully assigned kuttl-test-more-goldfish/with-badger-7b9bfb4788-96v7d to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:41 +0000 UTC Normal Pod with-badger-7b9bfb4788-96v7d AddedInterface Add eth0 [10.128.2.42/23] from ovn-kubernetes multus logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:41 +0000 UTC Normal Pod with-badger-7b9bfb4788-96v7d.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:41 +0000 UTC Normal Pod with-badger-7b9bfb4788-96v7d.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:41 +0000 UTC Normal Pod with-badger-7b9bfb4788-96v7d.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:41 +0000 UTC Normal Pod with-badger-7b9bfb4788-96v7d.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:41 +0000 UTC Normal Pod with-badger-7b9bfb4788-96v7d.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:41 +0000 UTC Normal Pod with-badger-7b9bfb4788-96v7d.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:41 +0000 UTC Normal ReplicaSet.apps with-badger-7b9bfb4788 SuccessfulCreate Created pod: with-badger-7b9bfb4788-96v7d replicaset-controller logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:41 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-7b9bfb4788 to 1 deployment-controller logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:42 +0000 UTC Normal Pod check-span-zzbgf Binding Scheduled Successfully assigned kuttl-test-more-goldfish/check-span-zzbgf to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:42 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-zzbgf job-controller logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:42 +0000 UTC Normal Pod report-span-tzvgb Binding Scheduled Successfully assigned kuttl-test-more-goldfish/report-span-tzvgb to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:42 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-tzvgb job-controller logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:43 +0000 UTC Normal Pod check-span-zzbgf AddedInterface Add eth0 [10.131.0.54/23] from ovn-kubernetes multus logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:43 +0000 UTC Normal Pod check-span-zzbgf.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:43 +0000 UTC Normal Pod report-span-tzvgb AddedInterface Add eth0 [10.129.2.40/23] from ovn-kubernetes multus logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:43 +0000 UTC Normal Pod report-span-tzvgb.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:44 +0000 UTC Normal Pod check-span-zzbgf.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 689ms (689ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:44 +0000 UTC Normal Pod check-span-zzbgf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:44 +0000 UTC Normal Pod check-span-zzbgf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:44 +0000 UTC Normal Pod report-span-tzvgb.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 637ms (637ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:44 +0000 UTC Normal Pod report-span-tzvgb.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:44 +0000 UTC Normal Pod report-span-tzvgb.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 20:30:56 | examples-with-badger | 2025-02-17 20:30:55 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 20:30:56 | examples-with-badger | Deleting namespace: kuttl-test-more-goldfish === CONT kuttl/harness/examples-simplest logger.go:42: 20:31:09 | examples-simplest | Creating namespace: kuttl-test-fine-grouper logger.go:42: 20:31:09 | examples-simplest/0-install | starting test step 0-install logger.go:42: 20:31:09 | examples-simplest/0-install | Jaeger:kuttl-test-fine-grouper/simplest created logger.go:42: 20:31:15 | examples-simplest/0-install | test step completed 0-install logger.go:42: 20:31:15 | examples-simplest/1-smoke-test | starting test step 1-smoke-test logger.go:42: 20:31:15 | examples-simplest/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 20:31:16 | examples-simplest/1-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:31:24 | examples-simplest/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 20:31:24 | examples-simplest/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 20:31:24 | examples-simplest/1-smoke-test | job.batch/report-span created logger.go:42: 20:31:25 | examples-simplest/1-smoke-test | job.batch/check-span created logger.go:42: 20:31:37 | examples-simplest/1-smoke-test | test step completed 1-smoke-test logger.go:42: 20:31:37 | examples-simplest | examples-simplest events from ns kuttl-test-fine-grouper: logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:13 +0000 UTC Normal Pod simplest-6fb554b5b8-rdfsr Binding Scheduled Successfully assigned kuttl-test-fine-grouper/simplest-6fb554b5b8-rdfsr to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:13 +0000 UTC Normal Pod simplest-6fb554b5b8-rdfsr AddedInterface Add eth0 [10.128.2.43/23] from ovn-kubernetes multus logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:13 +0000 UTC Normal Pod simplest-6fb554b5b8-rdfsr.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:13 +0000 UTC Normal Pod simplest-6fb554b5b8-rdfsr.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:13 +0000 UTC Normal Pod simplest-6fb554b5b8-rdfsr.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:13 +0000 UTC Normal Pod simplest-6fb554b5b8-rdfsr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:13 +0000 UTC Normal Pod simplest-6fb554b5b8-rdfsr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:13 +0000 UTC Normal Pod simplest-6fb554b5b8-rdfsr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:13 +0000 UTC Normal ReplicaSet.apps simplest-6fb554b5b8 SuccessfulCreate Created pod: simplest-6fb554b5b8-rdfsr replicaset-controller logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:13 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-6fb554b5b8 to 1 deployment-controller logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:21 +0000 UTC Normal Pod simplest-6fb554b5b8-rdfsr.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:21 +0000 UTC Normal Pod simplest-6fb554b5b8-rdfsr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:21 +0000 UTC Normal ReplicaSet.apps simplest-6fb554b5b8 SuccessfulDelete Deleted pod: simplest-6fb554b5b8-rdfsr replicaset-controller logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:21 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-6fb554b5b8 to 0 from 1 deployment-controller logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:22 +0000 UTC Normal Pod simplest-5994f98998-hjcqj Binding Scheduled Successfully assigned kuttl-test-fine-grouper/simplest-5994f98998-hjcqj to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:22 +0000 UTC Normal Pod simplest-5994f98998-hjcqj AddedInterface Add eth0 [10.128.2.44/23] from ovn-kubernetes multus logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:22 +0000 UTC Normal Pod simplest-5994f98998-hjcqj.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:22 +0000 UTC Normal ReplicaSet.apps simplest-5994f98998 SuccessfulCreate Created pod: simplest-5994f98998-hjcqj replicaset-controller logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:22 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-5994f98998 to 1 deployment-controller logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:23 +0000 UTC Normal Pod simplest-5994f98998-hjcqj.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:23 +0000 UTC Normal Pod simplest-5994f98998-hjcqj.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:23 +0000 UTC Normal Pod simplest-5994f98998-hjcqj.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:23 +0000 UTC Normal Pod simplest-5994f98998-hjcqj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:23 +0000 UTC Normal Pod simplest-5994f98998-hjcqj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:25 +0000 UTC Normal Pod check-span-kcc9x Binding Scheduled Successfully assigned kuttl-test-fine-grouper/check-span-kcc9x to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:25 +0000 UTC Normal Pod check-span-kcc9x AddedInterface Add eth0 [10.129.2.42/23] from ovn-kubernetes multus logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:25 +0000 UTC Normal Pod check-span-kcc9x.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:25 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-kcc9x job-controller logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:25 +0000 UTC Normal Pod report-span-l8xhw Binding Scheduled Successfully assigned kuttl-test-fine-grouper/report-span-l8xhw to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:25 +0000 UTC Normal Pod report-span-l8xhw AddedInterface Add eth0 [10.129.2.41/23] from ovn-kubernetes multus logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:25 +0000 UTC Normal Pod report-span-l8xhw.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:25 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-l8xhw job-controller logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:26 +0000 UTC Normal Pod check-span-kcc9x.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 778ms (778ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:26 +0000 UTC Normal Pod check-span-kcc9x.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:26 +0000 UTC Normal Pod check-span-kcc9x.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:26 +0000 UTC Normal Pod report-span-l8xhw.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.069s (1.069s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:26 +0000 UTC Normal Pod report-span-l8xhw.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:26 +0000 UTC Normal Pod report-span-l8xhw.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 20:31:37 | examples-simplest | 2025-02-17 20:31:36 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 20:31:37 | examples-simplest | Deleting namespace: kuttl-test-fine-grouper === CONT kuttl/harness/examples-simple-prod-with-volumes logger.go:42: 20:31:50 | examples-simple-prod-with-volumes | Ignoring 03-check-volume.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:31:50 | examples-simple-prod-with-volumes | Creating namespace: kuttl-test-infinite-seasnail logger.go:42: 20:31:50 | examples-simple-prod-with-volumes/1-install | starting test step 1-install logger.go:42: 20:31:50 | examples-simple-prod-with-volumes/1-install | Jaeger:kuttl-test-infinite-seasnail/simple-prod created logger.go:42: 20:32:27 | examples-simple-prod-with-volumes/1-install | test step completed 1-install logger.go:42: 20:32:27 | examples-simple-prod-with-volumes/2-smoke-test | starting test step 2-smoke-test logger.go:42: 20:32:27 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 20:32:28 | examples-simple-prod-with-volumes/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:32:34 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 20:32:35 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 20:32:35 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/report-span created logger.go:42: 20:32:35 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/check-span created logger.go:42: 20:32:47 | examples-simple-prod-with-volumes/2-smoke-test | test step completed 2-smoke-test logger.go:42: 20:32:47 | examples-simple-prod-with-volumes/3-check-volume | starting test step 3-check-volume logger.go:42: 20:32:47 | examples-simple-prod-with-volumes/3-check-volume | running command: [sh -c kubectl exec $(kubectl get pods -n $NAMESPACE -l app=jaeger -l app.kubernetes.io/component=collector -o yaml | /tmp/jaeger-tests/bin/yq e '.items[0].metadata.name') -n $NAMESPACE -- ls /usr/share/elasticsearch/data] logger.go:42: 20:32:48 | examples-simple-prod-with-volumes/3-check-volume | test step completed 3-check-volume logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | examples-simple-prod-with-volumes events from ns kuttl-test-infinite-seasnail: logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:31:58 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestinfiniteseasnailsimpleprod-1-7dd4cfff6d SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestinfiniteseasnailsimpleprod-1-7dw8bvk replicaset-controller logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:31:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinfiniteseasnailsimpleprod-1-7dw8bvk Binding Scheduled Successfully assigned kuttl-test-infinite-seasnail/elasticsearch-cdm-kuttltestinfiniteseasnailsimpleprod-1-7dw8bvk to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:31:58 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestinfiniteseasnailsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestinfiniteseasnailsimpleprod-1-7dd4cfff6d to 1 deployment-controller logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:31:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinfiniteseasnailsimpleprod-1-7dw8bvk AddedInterface Add eth0 [10.129.2.43/23] from ovn-kubernetes multus logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:31:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinfiniteseasnailsimpleprod-1-7dw8bvk.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:5117d0aee260f74d6ef77c7c71ed073b1ad45b456f58894539b5697bec878310" already present on machine kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:31:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinfiniteseasnailsimpleprod-1-7dw8bvk.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:31:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinfiniteseasnailsimpleprod-1-7dw8bvk.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:31:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinfiniteseasnailsimpleprod-1-7dw8bvk.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d63d612cdca72f64c45960a9b30bcfd0c746daabc17407df92c62694d5bc4bb6" already present on machine kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:31:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinfiniteseasnailsimpleprod-1-7dw8bvk.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:31:59 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinfiniteseasnailsimpleprod-1-7dw8bvk.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:14 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestinfiniteseasnailsimpleprod-1-7dw8bvk.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:24 +0000 UTC Normal Pod simple-prod-collector-76f7cd785d-q5jsq Binding Scheduled Successfully assigned kuttl-test-infinite-seasnail/simple-prod-collector-76f7cd785d-q5jsq to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:24 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-76f7cd785d SuccessfulCreate Created pod: simple-prod-collector-76f7cd785d-q5jsq replicaset-controller logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:24 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-76f7cd785d to 1 deployment-controller logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:24 +0000 UTC Normal Pod simple-prod-query-9f84d7ddf-rfcmp Binding Scheduled Successfully assigned kuttl-test-infinite-seasnail/simple-prod-query-9f84d7ddf-rfcmp to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:24 +0000 UTC Normal ReplicaSet.apps simple-prod-query-9f84d7ddf SuccessfulCreate Created pod: simple-prod-query-9f84d7ddf-rfcmp replicaset-controller logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:24 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-9f84d7ddf to 1 deployment-controller logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:25 +0000 UTC Normal Pod simple-prod-collector-76f7cd785d-q5jsq AddedInterface Add eth0 [10.131.0.55/23] from ovn-kubernetes multus logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:25 +0000 UTC Normal Pod simple-prod-collector-76f7cd785d-q5jsq.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:25 +0000 UTC Normal Pod simple-prod-collector-76f7cd785d-q5jsq.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:25 +0000 UTC Normal Pod simple-prod-collector-76f7cd785d-q5jsq.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:25 +0000 UTC Normal Pod simple-prod-query-9f84d7ddf-rfcmp AddedInterface Add eth0 [10.128.2.45/23] from ovn-kubernetes multus logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:25 +0000 UTC Normal Pod simple-prod-query-9f84d7ddf-rfcmp.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:25 +0000 UTC Normal Pod simple-prod-query-9f84d7ddf-rfcmp.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:25 +0000 UTC Normal Pod simple-prod-query-9f84d7ddf-rfcmp.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:25 +0000 UTC Normal Pod simple-prod-query-9f84d7ddf-rfcmp.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:25 +0000 UTC Normal Pod simple-prod-query-9f84d7ddf-rfcmp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:25 +0000 UTC Normal Pod simple-prod-query-9f84d7ddf-rfcmp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:25 +0000 UTC Normal Pod simple-prod-query-9f84d7ddf-rfcmp.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:25 +0000 UTC Normal Pod simple-prod-query-9f84d7ddf-rfcmp.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:25 +0000 UTC Normal Pod simple-prod-query-9f84d7ddf-rfcmp.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:31 +0000 UTC Normal Pod simple-prod-query-9f84d7ddf-rfcmp.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:31 +0000 UTC Normal Pod simple-prod-query-9f84d7ddf-rfcmp.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:31 +0000 UTC Normal Pod simple-prod-query-9f84d7ddf-rfcmp.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:31 +0000 UTC Normal ReplicaSet.apps simple-prod-query-9f84d7ddf SuccessfulDelete Deleted pod: simple-prod-query-9f84d7ddf-rfcmp replicaset-controller logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:31 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-9f84d7ddf to 0 from 1 deployment-controller logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:32 +0000 UTC Normal Pod simple-prod-query-c68df749b-nrpt8 Binding Scheduled Successfully assigned kuttl-test-infinite-seasnail/simple-prod-query-c68df749b-nrpt8 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:32 +0000 UTC Normal ReplicaSet.apps simple-prod-query-c68df749b SuccessfulCreate Created pod: simple-prod-query-c68df749b-nrpt8 replicaset-controller logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:32 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-c68df749b to 1 deployment-controller logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:33 +0000 UTC Normal Pod simple-prod-query-c68df749b-nrpt8 AddedInterface Add eth0 [10.128.2.46/23] from ovn-kubernetes multus logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:33 +0000 UTC Normal Pod simple-prod-query-c68df749b-nrpt8.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:33 +0000 UTC Normal Pod simple-prod-query-c68df749b-nrpt8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:33 +0000 UTC Normal Pod simple-prod-query-c68df749b-nrpt8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:33 +0000 UTC Normal Pod simple-prod-query-c68df749b-nrpt8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:33 +0000 UTC Normal Pod simple-prod-query-c68df749b-nrpt8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:33 +0000 UTC Normal Pod simple-prod-query-c68df749b-nrpt8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:33 +0000 UTC Normal Pod simple-prod-query-c68df749b-nrpt8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:33 +0000 UTC Normal Pod simple-prod-query-c68df749b-nrpt8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:33 +0000 UTC Normal Pod simple-prod-query-c68df749b-nrpt8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:35 +0000 UTC Normal Pod check-span-srtsn Binding Scheduled Successfully assigned kuttl-test-infinite-seasnail/check-span-srtsn to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:35 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-srtsn job-controller logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:35 +0000 UTC Normal Pod report-span-vnt5l Binding Scheduled Successfully assigned kuttl-test-infinite-seasnail/report-span-vnt5l to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:35 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-vnt5l job-controller logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:36 +0000 UTC Normal Pod check-span-srtsn AddedInterface Add eth0 [10.131.0.57/23] from ovn-kubernetes multus logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:36 +0000 UTC Normal Pod check-span-srtsn.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:36 +0000 UTC Normal Pod report-span-vnt5l AddedInterface Add eth0 [10.131.0.56/23] from ovn-kubernetes multus logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:36 +0000 UTC Normal Pod report-span-vnt5l.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:36 +0000 UTC Normal Pod report-span-vnt5l.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 761ms (761ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:36 +0000 UTC Normal Pod report-span-vnt5l.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:36 +0000 UTC Normal Pod report-span-vnt5l.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:37 +0000 UTC Normal Pod check-span-srtsn.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.012s (1.012s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:37 +0000 UTC Normal Pod check-span-srtsn.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:37 +0000 UTC Normal Pod check-span-srtsn.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | 2025-02-17 20:32:47 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 20:32:48 | examples-simple-prod-with-volumes | Deleting namespace: kuttl-test-infinite-seasnail === CONT kuttl/harness/examples-business-application-injected-sidecar logger.go:42: 20:32:56 | examples-business-application-injected-sidecar | Creating namespace: kuttl-test-infinite-ostrich logger.go:42: 20:32:56 | examples-business-application-injected-sidecar/0-install | starting test step 0-install logger.go:42: 20:32:56 | examples-business-application-injected-sidecar/0-install | Deployment:kuttl-test-infinite-ostrich/myapp created logger.go:42: 20:32:56 | examples-business-application-injected-sidecar/0-install | test step completed 0-install logger.go:42: 20:32:56 | examples-business-application-injected-sidecar/1-install | starting test step 1-install logger.go:42: 20:32:56 | examples-business-application-injected-sidecar/1-install | Jaeger:kuttl-test-infinite-ostrich/simplest created logger.go:42: 20:33:10 | examples-business-application-injected-sidecar/1-install | test step completed 1-install logger.go:42: 20:33:10 | examples-business-application-injected-sidecar/2-smoke-test | starting test step 2-smoke-test logger.go:42: 20:33:10 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 20:33:11 | examples-business-application-injected-sidecar/2-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:33:18 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 20:33:18 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 20:33:18 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/report-span created logger.go:42: 20:33:18 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/check-span created logger.go:42: 20:33:31 | examples-business-application-injected-sidecar/2-smoke-test | test step completed 2-smoke-test logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | examples-business-application-injected-sidecar events from ns kuttl-test-infinite-ostrich: logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:32:56 +0000 UTC Normal Pod myapp-68dd5ff844-6gb5c Binding Scheduled Successfully assigned kuttl-test-infinite-ostrich/myapp-68dd5ff844-6gb5c to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:32:56 +0000 UTC Warning Pod myapp-68dd5ff844-6gb5c FailedMount MountVolume.SetUp failed for volume "simplest-trusted-ca" : configmap "simplest-trusted-ca" not found kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:32:56 +0000 UTC Warning Pod myapp-68dd5ff844-6gb5c FailedMount MountVolume.SetUp failed for volume "simplest-service-ca" : configmap "simplest-service-ca" not found kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:32:56 +0000 UTC Normal ReplicaSet.apps myapp-68dd5ff844 SuccessfulCreate Created pod: myapp-68dd5ff844-6gb5c replicaset-controller logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:32:56 +0000 UTC Normal Pod myapp-fc54f78d8-hztsz Binding Scheduled Successfully assigned kuttl-test-infinite-ostrich/myapp-fc54f78d8-hztsz to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:32:56 +0000 UTC Normal Pod myapp-fc54f78d8-hztsz AddedInterface Add eth0 [10.129.2.44/23] from ovn-kubernetes multus logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:32:56 +0000 UTC Normal Pod myapp-fc54f78d8-hztsz.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:32:56 +0000 UTC Normal ReplicaSet.apps myapp-fc54f78d8 SuccessfulCreate Created pod: myapp-fc54f78d8-hztsz replicaset-controller logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:32:56 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-fc54f78d8 to 1 deployment-controller logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:32:56 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-68dd5ff844 to 1 deployment-controller logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:01 +0000 UTC Normal Pod myapp-fc54f78d8-hztsz.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 4.136s (4.136s including waiting). Image size: 282912835 bytes. kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:01 +0000 UTC Normal Pod myapp-fc54f78d8-hztsz.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:01 +0000 UTC Normal Pod myapp-fc54f78d8-hztsz.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:05 +0000 UTC Normal Pod simplest-76bb686b55-vjmt8 Binding Scheduled Successfully assigned kuttl-test-infinite-ostrich/simplest-76bb686b55-vjmt8 to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:05 +0000 UTC Normal ReplicaSet.apps simplest-76bb686b55 SuccessfulCreate Created pod: simplest-76bb686b55-vjmt8 replicaset-controller logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:05 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-76bb686b55 to 1 deployment-controller logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:06 +0000 UTC Normal Pod simplest-76bb686b55-vjmt8 AddedInterface Add eth0 [10.131.0.58/23] from ovn-kubernetes multus logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:06 +0000 UTC Normal Pod simplest-76bb686b55-vjmt8.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:07 +0000 UTC Warning Pod myapp-fc54f78d8-hztsz.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.129.2.44:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:08 +0000 UTC Normal Pod simplest-76bb686b55-vjmt8.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" in 2.298s (2.298s including waiting). Image size: 144304495 bytes. kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:08 +0000 UTC Normal Pod simplest-76bb686b55-vjmt8.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:08 +0000 UTC Normal Pod simplest-76bb686b55-vjmt8.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:08 +0000 UTC Normal Pod simplest-76bb686b55-vjmt8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:08 +0000 UTC Normal Pod simplest-76bb686b55-vjmt8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:08 +0000 UTC Normal Pod simplest-76bb686b55-vjmt8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:12 +0000 UTC Normal Pod myapp-68dd5ff844-6gb5c AddedInterface Add eth0 [10.128.2.47/23] from ovn-kubernetes multus logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:12 +0000 UTC Normal Pod myapp-68dd5ff844-6gb5c.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:14 +0000 UTC Normal Pod simplest-55696c84ff-lv7ww Binding Scheduled Successfully assigned kuttl-test-infinite-ostrich/simplest-55696c84ff-lv7ww to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:14 +0000 UTC Normal ReplicaSet.apps simplest-55696c84ff SuccessfulCreate Created pod: simplest-55696c84ff-lv7ww replicaset-controller logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:14 +0000 UTC Normal Pod simplest-76bb686b55-vjmt8.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:14 +0000 UTC Normal Pod simplest-76bb686b55-vjmt8.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:14 +0000 UTC Normal ReplicaSet.apps simplest-76bb686b55 SuccessfulDelete Deleted pod: simplest-76bb686b55-vjmt8 replicaset-controller logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:14 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-76bb686b55 to 0 from 1 deployment-controller logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:14 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-55696c84ff to 1 deployment-controller logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:15 +0000 UTC Normal Pod simplest-55696c84ff-lv7ww AddedInterface Add eth0 [10.131.0.59/23] from ovn-kubernetes multus logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:15 +0000 UTC Normal Pod simplest-55696c84ff-lv7ww.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:15 +0000 UTC Normal Pod simplest-55696c84ff-lv7ww.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:15 +0000 UTC Normal Pod simplest-55696c84ff-lv7ww.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:15 +0000 UTC Normal Pod simplest-55696c84ff-lv7ww.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:15 +0000 UTC Normal Pod simplest-55696c84ff-lv7ww.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:15 +0000 UTC Normal Pod simplest-55696c84ff-lv7ww.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:16 +0000 UTC Normal Pod myapp-68dd5ff844-6gb5c.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.649s (3.649s including waiting). Image size: 282912835 bytes. kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:16 +0000 UTC Normal Pod myapp-68dd5ff844-6gb5c.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:16 +0000 UTC Normal Pod myapp-68dd5ff844-6gb5c.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:16 +0000 UTC Normal Pod myapp-68dd5ff844-6gb5c.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:16 +0000 UTC Normal Pod myapp-68dd5ff844-6gb5c.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:16 +0000 UTC Normal Pod myapp-68dd5ff844-6gb5c.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:17 +0000 UTC Normal Pod myapp-fc54f78d8-hztsz.spec.containers{myapp} Killing Stopping container myapp kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:17 +0000 UTC Normal ReplicaSet.apps myapp-fc54f78d8 SuccessfulDelete Deleted pod: myapp-fc54f78d8-hztsz replicaset-controller logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:17 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled down replica set myapp-fc54f78d8 to 0 from 1 deployment-controller logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:18 +0000 UTC Normal Pod check-span-dvdvs Binding Scheduled Successfully assigned kuttl-test-infinite-ostrich/check-span-dvdvs to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:18 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-dvdvs job-controller logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:18 +0000 UTC Normal Pod report-span-8tlx8 Binding Scheduled Successfully assigned kuttl-test-infinite-ostrich/report-span-8tlx8 to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:18 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-8tlx8 job-controller logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:19 +0000 UTC Normal Pod check-span-dvdvs AddedInterface Add eth0 [10.129.2.46/23] from ovn-kubernetes multus logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:19 +0000 UTC Normal Pod check-span-dvdvs.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:19 +0000 UTC Normal Pod report-span-8tlx8 AddedInterface Add eth0 [10.129.2.45/23] from ovn-kubernetes multus logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:19 +0000 UTC Normal Pod report-span-8tlx8.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:20 +0000 UTC Normal Pod check-span-dvdvs.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 892ms (892ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:20 +0000 UTC Normal Pod check-span-dvdvs.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:20 +0000 UTC Normal Pod check-span-dvdvs.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:20 +0000 UTC Normal Pod report-span-8tlx8.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 586ms (586ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:20 +0000 UTC Normal Pod report-span-8tlx8.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:20 +0000 UTC Normal Pod report-span-8tlx8.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:23 +0000 UTC Warning Pod myapp-68dd5ff844-6gb5c.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.128.2.47:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | 2025-02-17 20:33:31 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 20:33:31 | examples-business-application-injected-sidecar | Deleting namespace: kuttl-test-infinite-ostrich === CONT kuttl/harness/examples-service-types logger.go:42: 20:33:44 | examples-service-types | Creating namespace: kuttl-test-liberal-puma logger.go:42: 20:33:44 | examples-service-types/0-install | starting test step 0-install logger.go:42: 20:33:44 | examples-service-types/0-install | Jaeger:kuttl-test-liberal-puma/service-types created logger.go:42: 20:33:51 | examples-service-types/0-install | test step completed 0-install logger.go:42: 20:33:51 | examples-service-types/1-smoke-test | starting test step 1-smoke-test logger.go:42: 20:33:51 | examples-service-types/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE service-types /dev/null] logger.go:42: 20:33:52 | examples-service-types/1-smoke-test | Warning: resource jaegers/service-types is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:33:59 | examples-service-types/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://service-types-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 20:33:59 | examples-service-types/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 20:33:59 | examples-service-types/1-smoke-test | job.batch/report-span created logger.go:42: 20:33:59 | examples-service-types/1-smoke-test | job.batch/check-span created logger.go:42: 20:34:13 | examples-service-types/1-smoke-test | test step completed 1-smoke-test logger.go:42: 20:34:13 | examples-service-types/2- | starting test step 2- logger.go:42: 20:34:13 | examples-service-types/2- | test step completed 2- logger.go:42: 20:34:13 | examples-service-types | examples-service-types events from ns kuttl-test-liberal-puma: logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:48 +0000 UTC Normal Pod service-types-5f8fdff7b9-c2hgx Binding Scheduled Successfully assigned kuttl-test-liberal-puma/service-types-5f8fdff7b9-c2hgx to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:48 +0000 UTC Normal ReplicaSet.apps service-types-5f8fdff7b9 SuccessfulCreate Created pod: service-types-5f8fdff7b9-c2hgx replicaset-controller logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:48 +0000 UTC Normal Service service-types-query EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:48 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-5f8fdff7b9 to 1 deployment-controller logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:49 +0000 UTC Normal Pod service-types-5f8fdff7b9-c2hgx AddedInterface Add eth0 [10.128.2.48/23] from ovn-kubernetes multus logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:49 +0000 UTC Normal Pod service-types-5f8fdff7b9-c2hgx.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:49 +0000 UTC Normal Pod service-types-5f8fdff7b9-c2hgx.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:49 +0000 UTC Normal Pod service-types-5f8fdff7b9-c2hgx.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:49 +0000 UTC Normal Pod service-types-5f8fdff7b9-c2hgx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:49 +0000 UTC Normal Pod service-types-5f8fdff7b9-c2hgx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:49 +0000 UTC Normal Pod service-types-5f8fdff7b9-c2hgx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:51 +0000 UTC Normal Service service-types-collector EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:51 +0000 UTC Normal Service service-types-query EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:53 +0000 UTC Normal Service service-types-collector EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:56 +0000 UTC Normal Pod service-types-5f8fdff7b9-c2hgx.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:56 +0000 UTC Normal Pod service-types-5f8fdff7b9-c2hgx.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:56 +0000 UTC Normal ReplicaSet.apps service-types-5f8fdff7b9 SuccessfulDelete Deleted pod: service-types-5f8fdff7b9-c2hgx replicaset-controller logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:56 +0000 UTC Normal Pod service-types-6894bb9c8f-x6jv2 Binding Scheduled Successfully assigned kuttl-test-liberal-puma/service-types-6894bb9c8f-x6jv2 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:56 +0000 UTC Normal ReplicaSet.apps service-types-6894bb9c8f SuccessfulCreate Created pod: service-types-6894bb9c8f-x6jv2 replicaset-controller logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:56 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled down replica set service-types-5f8fdff7b9 to 0 from 1 deployment-controller logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:56 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-6894bb9c8f to 1 deployment-controller logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:57 +0000 UTC Normal Pod service-types-6894bb9c8f-x6jv2 AddedInterface Add eth0 [10.128.2.49/23] from ovn-kubernetes multus logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:57 +0000 UTC Normal Pod service-types-6894bb9c8f-x6jv2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:57 +0000 UTC Normal Pod service-types-6894bb9c8f-x6jv2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:57 +0000 UTC Normal Pod service-types-6894bb9c8f-x6jv2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:57 +0000 UTC Normal Pod service-types-6894bb9c8f-x6jv2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:57 +0000 UTC Normal Pod service-types-6894bb9c8f-x6jv2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:57 +0000 UTC Normal Pod service-types-6894bb9c8f-x6jv2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:59 +0000 UTC Normal Pod check-span-r4lkz Binding Scheduled Successfully assigned kuttl-test-liberal-puma/check-span-r4lkz to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:59 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-r4lkz job-controller logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:59 +0000 UTC Normal Pod report-span-ndd79 Binding Scheduled Successfully assigned kuttl-test-liberal-puma/report-span-ndd79 to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:33:59 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-ndd79 job-controller logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:34:00 +0000 UTC Normal Pod check-span-r4lkz AddedInterface Add eth0 [10.129.2.48/23] from ovn-kubernetes multus logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:34:00 +0000 UTC Normal Pod check-span-r4lkz.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:34:00 +0000 UTC Normal Pod report-span-ndd79 AddedInterface Add eth0 [10.129.2.47/23] from ovn-kubernetes multus logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:34:00 +0000 UTC Normal Pod report-span-ndd79.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:34:01 +0000 UTC Normal Pod check-span-r4lkz.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.029s (1.029s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:34:01 +0000 UTC Normal Pod check-span-r4lkz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:34:01 +0000 UTC Normal Pod check-span-r4lkz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:34:01 +0000 UTC Normal Pod report-span-ndd79.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 602ms (602ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:34:01 +0000 UTC Normal Pod report-span-ndd79.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:34:01 +0000 UTC Normal Pod report-span-ndd79.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 20:34:13 | examples-service-types | 2025-02-17 20:34:12 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 20:34:13 | examples-service-types | Deleting namespace: kuttl-test-liberal-puma === CONT kuttl/harness/examples-openshift-with-htpasswd logger.go:42: 20:34:50 | examples-openshift-with-htpasswd | Ignoring 00-install.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:34:50 | examples-openshift-with-htpasswd | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:34:50 | examples-openshift-with-htpasswd | Creating namespace: kuttl-test-lasting-koala logger.go:42: 20:34:50 | examples-openshift-with-htpasswd/0-install | starting test step 0-install logger.go:42: 20:34:50 | examples-openshift-with-htpasswd/0-install | Secret:kuttl-test-lasting-koala/htpasswd created logger.go:42: 20:34:50 | examples-openshift-with-htpasswd/0-install | test step completed 0-install logger.go:42: 20:34:50 | examples-openshift-with-htpasswd/1-install | starting test step 1-install logger.go:42: 20:34:50 | examples-openshift-with-htpasswd/1-install | Jaeger:kuttl-test-lasting-koala/with-htpasswd created logger.go:42: 20:34:57 | examples-openshift-with-htpasswd/1-install | test step completed 1-install logger.go:42: 20:34:57 | examples-openshift-with-htpasswd/2-check-unsecured | starting test step 2-check-unsecured logger.go:42: 20:34:57 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [./ensure-ingress-host.sh] logger.go:42: 20:34:57 | examples-openshift-with-htpasswd/2-check-unsecured | Checking the Ingress host value was populated logger.go:42: 20:34:57 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 0 logger.go:42: 20:34:58 | examples-openshift-with-htpasswd/2-check-unsecured | Hostname is with-htpasswd-kuttl-test-lasting-koala.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com logger.go:42: 20:34:58 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 20:34:58 | examples-openshift-with-htpasswd/2-check-unsecured | Checking an expected HTTP response logger.go:42: 20:34:58 | examples-openshift-with-htpasswd/2-check-unsecured | Running in OpenShift logger.go:42: 20:34:58 | examples-openshift-with-htpasswd/2-check-unsecured | Not using any secret logger.go:42: 20:34:58 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 1/30 the https://with-htpasswd-kuttl-test-lasting-koala.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:34:58 | examples-openshift-with-htpasswd/2-check-unsecured | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 20:34:58 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 2/30 the https://with-htpasswd-kuttl-test-lasting-koala.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:34:58 | examples-openshift-with-htpasswd/2-check-unsecured | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 3/30 the https://with-htpasswd-kuttl-test-lasting-koala.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/2-check-unsecured | curl response asserted properly logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/2-check-unsecured | test step completed 2-check-unsecured logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/3-check-unauthorized | starting test step 3-check-unauthorized logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [./ensure-ingress-host.sh] logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking the Ingress host value was populated logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 0 logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/3-check-unauthorized | Hostname is with-htpasswd-kuttl-test-lasting-koala.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [sh -c JAEGER_USERNAME=wronguser JAEGER_PASSWORD=wrongpassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking an expected HTTP response logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/3-check-unauthorized | Running in OpenShift logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/3-check-unauthorized | Using Jaeger basic authentication logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 1/30 the https://with-htpasswd-kuttl-test-lasting-koala.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/3-check-unauthorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 2/30 the https://with-htpasswd-kuttl-test-lasting-koala.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/3-check-unauthorized | curl response asserted properly logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/3-check-unauthorized | test step completed 3-check-unauthorized logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/4-check-authorized | starting test step 4-check-authorized logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/4-check-authorized | running command: [./ensure-ingress-host.sh] logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/4-check-authorized | Checking the Ingress host value was populated logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/4-check-authorized | Try number 0 logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/4-check-authorized | Hostname is with-htpasswd-kuttl-test-lasting-koala.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com logger.go:42: 20:35:08 | examples-openshift-with-htpasswd/4-check-authorized | running command: [sh -c JAEGER_USERNAME=awesomeuser JAEGER_PASSWORD=awesomepassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE with-htpasswd] logger.go:42: 20:35:09 | examples-openshift-with-htpasswd/4-check-authorized | Checking an expected HTTP response logger.go:42: 20:35:09 | examples-openshift-with-htpasswd/4-check-authorized | Running in OpenShift logger.go:42: 20:35:09 | examples-openshift-with-htpasswd/4-check-authorized | Using Jaeger basic authentication logger.go:42: 20:35:09 | examples-openshift-with-htpasswd/4-check-authorized | Try number 1/30 the https://with-htpasswd-kuttl-test-lasting-koala.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:35:09 | examples-openshift-with-htpasswd/4-check-authorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 20:35:09 | examples-openshift-with-htpasswd/4-check-authorized | Try number 2/30 the https://with-htpasswd-kuttl-test-lasting-koala.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:35:09 | examples-openshift-with-htpasswd/4-check-authorized | curl response asserted properly logger.go:42: 20:35:09 | examples-openshift-with-htpasswd/4-check-authorized | test step completed 4-check-authorized logger.go:42: 20:35:09 | examples-openshift-with-htpasswd | examples-openshift-with-htpasswd events from ns kuttl-test-lasting-koala: logger.go:42: 20:35:09 | examples-openshift-with-htpasswd | 2025-02-17 20:34:54 +0000 UTC Normal Deployment.apps with-htpasswd ScalingReplicaSet Scaled up replica set with-htpasswd-d98f4dffd to 1 deployment-controller logger.go:42: 20:35:09 | examples-openshift-with-htpasswd | 2025-02-17 20:34:55 +0000 UTC Normal Pod with-htpasswd-d98f4dffd-zhvh2 Binding Scheduled Successfully assigned kuttl-test-lasting-koala/with-htpasswd-d98f4dffd-zhvh2 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:35:09 | examples-openshift-with-htpasswd | 2025-02-17 20:34:55 +0000 UTC Normal Pod with-htpasswd-d98f4dffd-zhvh2 AddedInterface Add eth0 [10.128.2.50/23] from ovn-kubernetes multus logger.go:42: 20:35:09 | examples-openshift-with-htpasswd | 2025-02-17 20:34:55 +0000 UTC Normal Pod with-htpasswd-d98f4dffd-zhvh2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:35:09 | examples-openshift-with-htpasswd | 2025-02-17 20:34:55 +0000 UTC Normal Pod with-htpasswd-d98f4dffd-zhvh2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:35:09 | examples-openshift-with-htpasswd | 2025-02-17 20:34:55 +0000 UTC Normal Pod with-htpasswd-d98f4dffd-zhvh2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:35:09 | examples-openshift-with-htpasswd | 2025-02-17 20:34:55 +0000 UTC Normal Pod with-htpasswd-d98f4dffd-zhvh2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:35:09 | examples-openshift-with-htpasswd | 2025-02-17 20:34:55 +0000 UTC Normal Pod with-htpasswd-d98f4dffd-zhvh2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:35:09 | examples-openshift-with-htpasswd | 2025-02-17 20:34:55 +0000 UTC Normal Pod with-htpasswd-d98f4dffd-zhvh2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:35:09 | examples-openshift-with-htpasswd | 2025-02-17 20:34:55 +0000 UTC Normal ReplicaSet.apps with-htpasswd-d98f4dffd SuccessfulCreate Created pod: with-htpasswd-d98f4dffd-zhvh2 replicaset-controller logger.go:42: 20:35:09 | examples-openshift-with-htpasswd | Deleting namespace: kuttl-test-lasting-koala === CONT kuttl/harness/examples-all-in-one-with-options logger.go:42: 20:35:15 | examples-all-in-one-with-options | Creating namespace: kuttl-test-humorous-pipefish logger.go:42: 20:35:15 | examples-all-in-one-with-options/0-install | starting test step 0-install logger.go:42: 20:35:15 | examples-all-in-one-with-options/0-install | Jaeger:kuttl-test-humorous-pipefish/my-jaeger created logger.go:42: 20:35:22 | examples-all-in-one-with-options/0-install | test step completed 0-install logger.go:42: 20:35:22 | examples-all-in-one-with-options/1-smoke-test | starting test step 1-smoke-test logger.go:42: 20:35:22 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 20:35:24 | examples-all-in-one-with-options/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:35:30 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443/jaeger MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 20:35:31 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 20:35:31 | examples-all-in-one-with-options/1-smoke-test | job.batch/report-span created logger.go:42: 20:35:31 | examples-all-in-one-with-options/1-smoke-test | job.batch/check-span created logger.go:42: 20:35:43 | examples-all-in-one-with-options/1-smoke-test | test step completed 1-smoke-test logger.go:42: 20:35:43 | examples-all-in-one-with-options | examples-all-in-one-with-options events from ns kuttl-test-humorous-pipefish: logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:19 +0000 UTC Normal Pod my-jaeger-6564d4ff8f-gf5d2 Binding Scheduled Successfully assigned kuttl-test-humorous-pipefish/my-jaeger-6564d4ff8f-gf5d2 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:19 +0000 UTC Normal ReplicaSet.apps my-jaeger-6564d4ff8f SuccessfulCreate Created pod: my-jaeger-6564d4ff8f-gf5d2 replicaset-controller logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:19 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6564d4ff8f to 1 deployment-controller logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:20 +0000 UTC Normal Pod my-jaeger-6564d4ff8f-gf5d2 AddedInterface Add eth0 [10.128.2.51/23] from ovn-kubernetes multus logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:20 +0000 UTC Normal Pod my-jaeger-6564d4ff8f-gf5d2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:20 +0000 UTC Normal Pod my-jaeger-6564d4ff8f-gf5d2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:20 +0000 UTC Normal Pod my-jaeger-6564d4ff8f-gf5d2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:20 +0000 UTC Normal Pod my-jaeger-6564d4ff8f-gf5d2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:20 +0000 UTC Normal Pod my-jaeger-6564d4ff8f-gf5d2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:20 +0000 UTC Normal Pod my-jaeger-6564d4ff8f-gf5d2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:26 +0000 UTC Normal Pod my-jaeger-6564d4ff8f-gf5d2.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:26 +0000 UTC Normal Pod my-jaeger-6564d4ff8f-gf5d2.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:26 +0000 UTC Normal ReplicaSet.apps my-jaeger-6564d4ff8f SuccessfulDelete Deleted pod: my-jaeger-6564d4ff8f-gf5d2 replicaset-controller logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:26 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-6564d4ff8f to 0 from 1 deployment-controller logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:26 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-798f4f5bd9 to 1 deployment-controller logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:27 +0000 UTC Normal Pod my-jaeger-798f4f5bd9-5tndn Binding Scheduled Successfully assigned kuttl-test-humorous-pipefish/my-jaeger-798f4f5bd9-5tndn to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:27 +0000 UTC Normal Pod my-jaeger-798f4f5bd9-5tndn AddedInterface Add eth0 [10.128.2.52/23] from ovn-kubernetes multus logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:27 +0000 UTC Normal Pod my-jaeger-798f4f5bd9-5tndn.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:27 +0000 UTC Normal Pod my-jaeger-798f4f5bd9-5tndn.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:27 +0000 UTC Normal Pod my-jaeger-798f4f5bd9-5tndn.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:27 +0000 UTC Normal Pod my-jaeger-798f4f5bd9-5tndn.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:27 +0000 UTC Normal Pod my-jaeger-798f4f5bd9-5tndn.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:27 +0000 UTC Normal Pod my-jaeger-798f4f5bd9-5tndn.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:27 +0000 UTC Normal ReplicaSet.apps my-jaeger-798f4f5bd9 SuccessfulCreate Created pod: my-jaeger-798f4f5bd9-5tndn replicaset-controller logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:31 +0000 UTC Normal Pod check-span-szfx7 Binding Scheduled Successfully assigned kuttl-test-humorous-pipefish/check-span-szfx7 to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:31 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-szfx7 job-controller logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:31 +0000 UTC Normal Pod report-span-jj9mf Binding Scheduled Successfully assigned kuttl-test-humorous-pipefish/report-span-jj9mf to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:31 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-jj9mf job-controller logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:32 +0000 UTC Normal Pod check-span-szfx7 AddedInterface Add eth0 [10.129.2.50/23] from ovn-kubernetes multus logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:32 +0000 UTC Normal Pod check-span-szfx7.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:32 +0000 UTC Normal Pod check-span-szfx7.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 578ms (578ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:32 +0000 UTC Normal Pod check-span-szfx7.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:32 +0000 UTC Normal Pod check-span-szfx7.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:32 +0000 UTC Normal Pod report-span-jj9mf AddedInterface Add eth0 [10.129.2.49/23] from ovn-kubernetes multus logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:32 +0000 UTC Normal Pod report-span-jj9mf.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:32 +0000 UTC Normal Pod report-span-jj9mf.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 823ms (823ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:32 +0000 UTC Normal Pod report-span-jj9mf.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:32 +0000 UTC Normal Pod report-span-jj9mf.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 20:35:43 | examples-all-in-one-with-options | 2025-02-17 20:35:43 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 20:35:43 | examples-all-in-one-with-options | Deleting namespace: kuttl-test-humorous-pipefish === CONT kuttl/harness/examples-agent-with-priority-class logger.go:42: 20:35:56 | examples-agent-with-priority-class | Creating namespace: kuttl-test-notable-primate logger.go:42: 20:35:56 | examples-agent-with-priority-class/0-install | starting test step 0-install logger.go:42: 20:35:56 | examples-agent-with-priority-class/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 20:35:56 | examples-agent-with-priority-class/0-install | ServiceAccount:kuttl-test-notable-primate/jaeger-agent-daemonset created logger.go:42: 20:35:56 | examples-agent-with-priority-class/0-install | test step completed 0-install logger.go:42: 20:35:56 | examples-agent-with-priority-class/1-add-policy | starting test step 1-add-policy logger.go:42: 20:35:56 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 20:35:56 | examples-agent-with-priority-class/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 20:35:56 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 20:36:01 | examples-agent-with-priority-class/1-add-policy | test step completed 1-add-policy logger.go:42: 20:36:01 | examples-agent-with-priority-class/2-install | starting test step 2-install logger.go:42: 20:36:01 | examples-agent-with-priority-class/2-install | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 20:36:08 | examples-agent-with-priority-class/2-install | Error from server (NotFound): deployments.apps "agent-as-daemonset" not found logger.go:42: 20:36:08 | examples-agent-with-priority-class/2-install | command failure, skipping 2 additional commands logger.go:42: 20:36:09 | examples-agent-with-priority-class/2-install | PriorityClass:/high-priority created logger.go:42: 20:36:09 | examples-agent-with-priority-class/2-install | Jaeger:kuttl-test-notable-primate/agent-as-daemonset updated case.go:364: failed in step 2-install case.go:366: exit status 1 logger.go:42: 20:36:09 | examples-agent-with-priority-class | examples-agent-with-priority-class events from ns kuttl-test-notable-primate: logger.go:42: 20:36:09 | examples-agent-with-priority-class | Deleting namespace: kuttl-test-notable-primate === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (542.42s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.10s) --- PASS: kuttl/harness/examples-simple-prod (72.57s) --- PASS: kuttl/harness/examples-with-sampling (59.76s) --- PASS: kuttl/harness/examples-with-cassandra (56.20s) --- PASS: kuttl/harness/examples-with-badger (41.08s) --- PASS: kuttl/harness/examples-simplest (41.07s) --- PASS: kuttl/harness/examples-simple-prod-with-volumes (65.93s) --- PASS: kuttl/harness/examples-business-application-injected-sidecar (48.24s) --- PASS: kuttl/harness/examples-service-types (66.31s) --- PASS: kuttl/harness/examples-openshift-with-htpasswd (25.14s) --- PASS: kuttl/harness/examples-all-in-one-with-options (40.53s) --- FAIL: kuttl/harness/examples-agent-with-priority-class (19.36s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name examples --report --output /logs/artifacts/examples.xml ./artifacts/kuttl-report.xml time="2025-02-17T20:36:15Z" level=debug msg="Setting a new name for the test suites" time="2025-02-17T20:36:15Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-02-17T20:36:15Z" level=debug msg="normalizing test case names" time="2025-02-17T20:36:15Z" level=debug msg="examples/artifacts -> examples_artifacts" time="2025-02-17T20:36:15Z" level=debug msg="examples/examples-simple-prod -> examples_examples_simple_prod" time="2025-02-17T20:36:15Z" level=debug msg="examples/examples-with-sampling -> examples_examples_with_sampling" time="2025-02-17T20:36:15Z" level=debug msg="examples/examples-with-cassandra -> examples_examples_with_cassandra" time="2025-02-17T20:36:15Z" level=debug msg="examples/examples-with-badger -> examples_examples_with_badger" time="2025-02-17T20:36:15Z" level=debug msg="examples/examples-simplest -> examples_examples_simplest" time="2025-02-17T20:36:15Z" level=debug msg="examples/examples-simple-prod-with-volumes -> examples_examples_simple_prod_with_volumes" time="2025-02-17T20:36:15Z" level=debug msg="examples/examples-business-application-injected-sidecar -> examples_examples_business_application_injected_sidecar" time="2025-02-17T20:36:15Z" level=debug msg="examples/examples-service-types -> examples_examples_service_types" time="2025-02-17T20:36:15Z" level=debug msg="examples/examples-openshift-with-htpasswd -> examples_examples_openshift_with_htpasswd" time="2025-02-17T20:36:15Z" level=debug msg="examples/examples-all-in-one-with-options -> examples_examples_all_in_one_with_options" time="2025-02-17T20:36:15Z" level=debug msg="examples/examples-agent-with-priority-class -> examples_examples_agent_with_priority_class" +---------------------------------------------------------+--------+ | NAME | RESULT | +---------------------------------------------------------+--------+ | examples_artifacts | passed | | examples_examples_simple_prod | passed | | examples_examples_with_sampling | passed | | examples_examples_with_cassandra | passed | | examples_examples_with_badger | passed | | examples_examples_simplest | passed | | examples_examples_simple_prod_with_volumes | passed | | examples_examples_business_application_injected_sidecar | passed | | examples_examples_service_types | passed | | examples_examples_openshift_with_htpasswd | passed | | examples_examples_all_in_one_with_options | passed | | examples_examples_agent_with_priority_class | failed | +---------------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh generate false true + '[' 3 -ne 3 ']' + test_suite_name=generate + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/generate.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-generate make[2]: Entering directory '/tmp/jaeger-tests' test -s /tmp/jaeger-tests/bin/operator-sdk || curl -sLo /tmp/jaeger-tests/bin/operator-sdk https://github.com/operator-framework/operator-sdk/releases/download/v1.32.0/operator-sdk_`go env GOOS`_`go env GOARCH` ./hack/install/install-golangci-lint.sh Installing golangci-lint Try 0... go install github.com/golangci/golangci-lint/cmd/golangci-lint@v1.55.2 ./hack/install/install-goimports.sh Installing goimports Try 0... go install golang.org/x/tools/cmd/goimports@v0.1.12 >>>> Formatting code... ./.ci/format.sh >>>> Building... ./hack/install/install-dependencies.sh Installing go dependencies Try 0... go mod download GOOS= GOARCH= CGO_ENABLED=0 GO111MODULE=on go build -ldflags "-X "github.com/jaegertracing/jaeger-operator/pkg/version".version="1.62.0" -X "github.com/jaegertracing/jaeger-operator/pkg/version".buildDate=2025-02-17T20:36:19Z -X "github.com/jaegertracing/jaeger-operator/pkg/version".defaultJaeger="1.62.0"" -o "bin/jaeger-operator" main.go JAEGER_VERSION="1.62.0" ./tests/e2e/generate/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-14-222249 True False 32m Cluster version is 4.18.0-0.nightly-2025-02-14-222249' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-14-222249 True False 32m Cluster version is 4.18.0-0.nightly-2025-02-14-222249' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/generate/render.sh ++ export SUITE_DIR=./tests/e2e/generate ++ SUITE_DIR=./tests/e2e/generate ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/generate ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test generate 'This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 2 -ne 2 ']' + test_name=generate + message='This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/generate/_build + '[' _build '!=' _build ']' + rm -rf generate + warning 'generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed\e[0m' WAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running generate E2E tests' Running generate E2E tests + cd tests/e2e/generate/_build + set +e + KUBECONFIG=/tmp/kubeconfig-998185858 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 20:36:36 | artifacts | Creating namespace: kuttl-test-stirring-skylark logger.go:42: 20:36:36 | artifacts | artifacts events from ns kuttl-test-stirring-skylark: logger.go:42: 20:36:36 | artifacts | Deleting namespace: kuttl-test-stirring-skylark === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.68s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.50s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name generate --report --output /logs/artifacts/generate.xml ./artifacts/kuttl-report.xml time="2025-02-17T20:36:43Z" level=debug msg="Setting a new name for the test suites" time="2025-02-17T20:36:43Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-02-17T20:36:43Z" level=debug msg="normalizing test case names" time="2025-02-17T20:36:43Z" level=debug msg="generate/artifacts -> generate_artifacts" +--------------------+--------+ | NAME | RESULT | +--------------------+--------+ | generate_artifacts | passed | +--------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh miscellaneous false true + '[' 3 -ne 3 ']' + test_suite_name=miscellaneous + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/miscellaneous.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-miscellaneous make[2]: Entering directory '/tmp/jaeger-tests' SKIP_ES_EXTERNAL=true ./tests/e2e/miscellaneous/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-14-222249 True False 33m Cluster version is 4.18.0-0.nightly-2025-02-14-222249' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-14-222249 True False 33m Cluster version is 4.18.0-0.nightly-2025-02-14-222249' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/miscellaneous/render.sh ++ export SUITE_DIR=./tests/e2e/miscellaneous ++ SUITE_DIR=./tests/e2e/miscellaneous ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/miscellaneous ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test collector-autoscale + '[' 1 -ne 1 ']' + test_name=collector-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-autoscale\e[0m' Rendering files for test collector-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p collector-autoscale + cd collector-autoscale + jaeger_name=simple-prod + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + ELASTICSEARCH_NODECOUNT=1 + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.resources.requests.memory="200m"' 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.autoscale=true 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.minReplicas=1 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.maxReplicas=2 01-install.yaml + version_lt 1.30 1.23 ++ echo 1.30 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.30 '!=' 1.30 + rm ./03-assert.yaml + generate_otlp_e2e_tests http + test_protocol=http + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-http\e[0m' Rendering files for test collector-otlp-allinone-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-autoscale + '[' collector-autoscale '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-http + cd collector-otlp-allinone-http + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger http true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-http\e[0m' Rendering files for test collector-otlp-production-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-http + '[' collector-otlp-allinone-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-http + cd collector-otlp-production-http + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger http true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + generate_otlp_e2e_tests grpc + test_protocol=grpc + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-grpc\e[0m' Rendering files for test collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-http + '[' collector-otlp-production-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-grpc + cd collector-otlp-allinone-grpc + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-grpc\e[0m' Rendering files for test collector-otlp-production-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-grpc + '[' collector-otlp-allinone-grpc '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-grpc + cd collector-otlp-production-grpc + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + '[' true = true ']' + skip_test istio 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=istio + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-grpc + '[' collector-otlp-production-grpc '!=' _build ']' + cd .. + rm -rf istio + warning 'istio: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: istio: Test not supported in OpenShift\e[0m' WAR: istio: Test not supported in OpenShift + '[' true = true ']' + skip_test outside-cluster 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=outside-cluster + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf outside-cluster + warning 'outside-cluster: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: outside-cluster: Test not supported in OpenShift\e[0m' WAR: outside-cluster: Test not supported in OpenShift + start_test set-custom-img + '[' 1 -ne 1 ']' + test_name=set-custom-img + echo =========================================================================== =========================================================================== + info 'Rendering files for test set-custom-img' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test set-custom-img\e[0m' Rendering files for test set-custom-img + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p set-custom-img + cd set-custom-img + jaeger_name=my-jaeger + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.image="test"' ./02-install.yaml + '[' true = true ']' + skip_test non-cluster-wide 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=non-cluster-wide + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/set-custom-img + '[' set-custom-img '!=' _build ']' + cd .. + rm -rf non-cluster-wide + warning 'non-cluster-wide: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: non-cluster-wide: Test not supported in OpenShift\e[0m' WAR: non-cluster-wide: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running miscellaneous E2E tests' Running miscellaneous E2E tests + cd tests/e2e/miscellaneous/_build + set +e + KUBECONFIG=/tmp/kubeconfig-998185858 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 8 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/cassandra-spark === PAUSE kuttl/harness/cassandra-spark === RUN kuttl/harness/collector-autoscale === PAUSE kuttl/harness/collector-autoscale === RUN kuttl/harness/collector-otlp-allinone-grpc === PAUSE kuttl/harness/collector-otlp-allinone-grpc === RUN kuttl/harness/collector-otlp-allinone-http === PAUSE kuttl/harness/collector-otlp-allinone-http === RUN kuttl/harness/collector-otlp-production-grpc === PAUSE kuttl/harness/collector-otlp-production-grpc === RUN kuttl/harness/collector-otlp-production-http === PAUSE kuttl/harness/collector-otlp-production-http === RUN kuttl/harness/set-custom-img === PAUSE kuttl/harness/set-custom-img === CONT kuttl/harness/artifacts logger.go:42: 20:36:54 | artifacts | Creating namespace: kuttl-test-arriving-kingfish logger.go:42: 20:36:54 | artifacts | artifacts events from ns kuttl-test-arriving-kingfish: logger.go:42: 20:36:54 | artifacts | Deleting namespace: kuttl-test-arriving-kingfish === CONT kuttl/harness/collector-otlp-allinone-http logger.go:42: 20:37:01 | collector-otlp-allinone-http | Creating namespace: kuttl-test-dashing-doberman logger.go:42: 20:37:01 | collector-otlp-allinone-http/0-install | starting test step 0-install logger.go:42: 20:37:01 | collector-otlp-allinone-http/0-install | Jaeger:kuttl-test-dashing-doberman/my-jaeger created logger.go:42: 20:37:07 | collector-otlp-allinone-http/0-install | test step completed 0-install logger.go:42: 20:37:07 | collector-otlp-allinone-http/1-smoke-test | starting test step 1-smoke-test logger.go:42: 20:37:07 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 20:37:08 | collector-otlp-allinone-http/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:37:15 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 20:37:15 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 20:37:16 | collector-otlp-allinone-http/1-smoke-test | job.batch/report-span created logger.go:42: 20:37:16 | collector-otlp-allinone-http/1-smoke-test | job.batch/check-span created logger.go:42: 20:37:28 | collector-otlp-allinone-http/1-smoke-test | test step completed 1-smoke-test logger.go:42: 20:37:28 | collector-otlp-allinone-http | collector-otlp-allinone-http events from ns kuttl-test-dashing-doberman: logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:05 +0000 UTC Normal Pod my-jaeger-56fbc8559b-ngj9f Binding Scheduled Successfully assigned kuttl-test-dashing-doberman/my-jaeger-56fbc8559b-ngj9f to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:05 +0000 UTC Normal ReplicaSet.apps my-jaeger-56fbc8559b SuccessfulCreate Created pod: my-jaeger-56fbc8559b-ngj9f replicaset-controller logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:05 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-56fbc8559b to 1 deployment-controller logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:06 +0000 UTC Normal Pod my-jaeger-56fbc8559b-ngj9f AddedInterface Add eth0 [10.128.2.53/23] from ovn-kubernetes multus logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:06 +0000 UTC Normal Pod my-jaeger-56fbc8559b-ngj9f.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:06 +0000 UTC Normal Pod my-jaeger-56fbc8559b-ngj9f.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:06 +0000 UTC Normal Pod my-jaeger-56fbc8559b-ngj9f.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:06 +0000 UTC Normal Pod my-jaeger-56fbc8559b-ngj9f.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:06 +0000 UTC Normal Pod my-jaeger-56fbc8559b-ngj9f.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:06 +0000 UTC Normal Pod my-jaeger-56fbc8559b-ngj9f.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:10 +0000 UTC Normal Pod my-jaeger-56fbc8559b-ngj9f.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:10 +0000 UTC Normal Pod my-jaeger-56fbc8559b-ngj9f.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:10 +0000 UTC Normal ReplicaSet.apps my-jaeger-56fbc8559b SuccessfulDelete Deleted pod: my-jaeger-56fbc8559b-ngj9f replicaset-controller logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:10 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-56fbc8559b to 0 from 1 deployment-controller logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:11 +0000 UTC Normal Pod my-jaeger-f98fcd7cd-h59fv Binding Scheduled Successfully assigned kuttl-test-dashing-doberman/my-jaeger-f98fcd7cd-h59fv to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:11 +0000 UTC Normal Pod my-jaeger-f98fcd7cd-h59fv AddedInterface Add eth0 [10.128.2.54/23] from ovn-kubernetes multus logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:11 +0000 UTC Normal ReplicaSet.apps my-jaeger-f98fcd7cd SuccessfulCreate Created pod: my-jaeger-f98fcd7cd-h59fv replicaset-controller logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:11 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-f98fcd7cd to 1 deployment-controller logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:12 +0000 UTC Normal Pod my-jaeger-f98fcd7cd-h59fv.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:12 +0000 UTC Normal Pod my-jaeger-f98fcd7cd-h59fv.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:12 +0000 UTC Normal Pod my-jaeger-f98fcd7cd-h59fv.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:12 +0000 UTC Normal Pod my-jaeger-f98fcd7cd-h59fv.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:12 +0000 UTC Normal Pod my-jaeger-f98fcd7cd-h59fv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:12 +0000 UTC Normal Pod my-jaeger-f98fcd7cd-h59fv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:16 +0000 UTC Normal Pod check-span-bsw89 Binding Scheduled Successfully assigned kuttl-test-dashing-doberman/check-span-bsw89 to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:16 +0000 UTC Normal Pod check-span-bsw89 AddedInterface Add eth0 [10.129.2.52/23] from ovn-kubernetes multus logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:16 +0000 UTC Normal Pod check-span-bsw89.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:16 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-bsw89 job-controller logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:16 +0000 UTC Normal Pod report-span-ph7cv Binding Scheduled Successfully assigned kuttl-test-dashing-doberman/report-span-ph7cv to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:16 +0000 UTC Normal Pod report-span-ph7cv AddedInterface Add eth0 [10.129.2.51/23] from ovn-kubernetes multus logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:16 +0000 UTC Normal Pod report-span-ph7cv.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:16 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-ph7cv job-controller logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:17 +0000 UTC Normal Pod check-span-bsw89.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 684ms (684ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:17 +0000 UTC Normal Pod check-span-bsw89.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:17 +0000 UTC Normal Pod check-span-bsw89.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:17 +0000 UTC Normal Pod report-span-ph7cv.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 575ms (575ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:17 +0000 UTC Normal Pod report-span-ph7cv.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:17 +0000 UTC Normal Pod report-span-ph7cv.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 20:37:28 | collector-otlp-allinone-http | 2025-02-17 20:37:27 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 20:37:28 | collector-otlp-allinone-http | Deleting namespace: kuttl-test-dashing-doberman === CONT kuttl/harness/set-custom-img logger.go:42: 20:37:41 | set-custom-img | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:37:41 | set-custom-img | Ignoring check-collector-img.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:37:41 | set-custom-img | Creating namespace: kuttl-test-perfect-monitor logger.go:42: 20:37:41 | set-custom-img/1-install | starting test step 1-install logger.go:42: 20:37:41 | set-custom-img/1-install | Jaeger:kuttl-test-perfect-monitor/my-jaeger created logger.go:42: 20:38:18 | set-custom-img/1-install | test step completed 1-install logger.go:42: 20:38:18 | set-custom-img/2-install | starting test step 2-install logger.go:42: 20:38:18 | set-custom-img/2-install | Jaeger:kuttl-test-perfect-monitor/my-jaeger updated logger.go:42: 20:38:18 | set-custom-img/2-install | test step completed 2-install logger.go:42: 20:38:18 | set-custom-img/3-check-image | starting test step 3-check-image logger.go:42: 20:38:18 | set-custom-img/3-check-image | running command: [sh -c ./check-collector-img.sh] logger.go:42: 20:38:18 | set-custom-img/3-check-image | Collector image missmatch. Expected: test. Has: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1 logger.go:42: 20:38:23 | set-custom-img/3-check-image | Collector image asserted properly! logger.go:42: 20:38:23 | set-custom-img/3-check-image | test step completed 3-check-image logger.go:42: 20:38:23 | set-custom-img | set-custom-img events from ns kuttl-test-perfect-monitor: logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:37:47 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestperfectmonitormyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestperfectmonitormyjaeger-1-6d54bc95d to 1 deployment-controller logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:37:48 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestperfectmonitormyjaeger-1-6d54bc95d SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestperfectmonitormyjaeger-1-6d54bc9jwr9 replicaset-controller logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:37:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestperfectmonitormyjaeger-1-6d54bc9jwr9 Binding Scheduled Successfully assigned kuttl-test-perfect-monitor/elasticsearch-cdm-kuttltestperfectmonitormyjaeger-1-6d54bc9jwr9 to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:37:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestperfectmonitormyjaeger-1-6d54bc9jwr9 AddedInterface Add eth0 [10.129.2.53/23] from ovn-kubernetes multus logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:37:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestperfectmonitormyjaeger-1-6d54bc9jwr9.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:5117d0aee260f74d6ef77c7c71ed073b1ad45b456f58894539b5697bec878310" already present on machine kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:37:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestperfectmonitormyjaeger-1-6d54bc9jwr9.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:37:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestperfectmonitormyjaeger-1-6d54bc9jwr9.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:37:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestperfectmonitormyjaeger-1-6d54bc9jwr9.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d63d612cdca72f64c45960a9b30bcfd0c746daabc17407df92c62694d5bc4bb6" already present on machine kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:37:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestperfectmonitormyjaeger-1-6d54bc9jwr9.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:37:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestperfectmonitormyjaeger-1-6d54bc9jwr9.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:03 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestperfectmonitormyjaeger-1-6d54bc9jwr9.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:14 +0000 UTC Normal Pod my-jaeger-collector-7784957795-z9kjk Binding Scheduled Successfully assigned kuttl-test-perfect-monitor/my-jaeger-collector-7784957795-z9kjk to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:14 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7784957795 SuccessfulCreate Created pod: my-jaeger-collector-7784957795-z9kjk replicaset-controller logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:14 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7784957795 to 1 deployment-controller logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:14 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-6f75cdb648 SuccessfulCreate Created pod: my-jaeger-query-6f75cdb648-ndp8k replicaset-controller logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:14 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-6f75cdb648 to 1 deployment-controller logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:15 +0000 UTC Normal Pod my-jaeger-query-6f75cdb648-ndp8k Binding Scheduled Successfully assigned kuttl-test-perfect-monitor/my-jaeger-query-6f75cdb648-ndp8k to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:15 +0000 UTC Normal Pod my-jaeger-query-6f75cdb648-ndp8k AddedInterface Add eth0 [10.131.0.60/23] from ovn-kubernetes multus logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:15 +0000 UTC Normal Pod my-jaeger-query-6f75cdb648-ndp8k.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:15 +0000 UTC Normal Pod my-jaeger-query-6f75cdb648-ndp8k.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:15 +0000 UTC Normal Pod my-jaeger-query-6f75cdb648-ndp8k.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:15 +0000 UTC Normal Pod my-jaeger-query-6f75cdb648-ndp8k.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:15 +0000 UTC Normal Pod my-jaeger-query-6f75cdb648-ndp8k.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:15 +0000 UTC Normal Pod my-jaeger-query-6f75cdb648-ndp8k.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:15 +0000 UTC Normal Pod my-jaeger-query-6f75cdb648-ndp8k.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:15 +0000 UTC Normal Pod my-jaeger-query-6f75cdb648-ndp8k.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:15 +0000 UTC Normal Pod my-jaeger-query-6f75cdb648-ndp8k.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:16 +0000 UTC Warning Pod my-jaeger-collector-7784957795-z9kjk FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:16 +0000 UTC Warning Pod my-jaeger-collector-7784957795-z9kjk FailedMount MountVolume.SetUp failed for volume "my-jaeger-trusted-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:16 +0000 UTC Normal Pod my-jaeger-collector-7784957795-z9kjk AddedInterface Add eth0 [10.128.2.55/23] from ovn-kubernetes multus logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:17 +0000 UTC Normal Pod my-jaeger-collector-7784957795-z9kjk.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:17 +0000 UTC Normal Pod my-jaeger-collector-7784957795-z9kjk.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:17 +0000 UTC Normal Pod my-jaeger-collector-7784957795-z9kjk.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:20 +0000 UTC Normal Pod my-jaeger-collector-7784957795-z9kjk.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:20 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7784957795 SuccessfulDelete Deleted pod: my-jaeger-collector-7784957795-z9kjk replicaset-controller logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:20 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-7784957795 to 0 from 1 deployment-controller logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:21 +0000 UTC Normal Pod my-jaeger-collector-598dc889b8-qbkpn Binding Scheduled Successfully assigned kuttl-test-perfect-monitor/my-jaeger-collector-598dc889b8-qbkpn to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:21 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-598dc889b8 SuccessfulCreate Created pod: my-jaeger-collector-598dc889b8-qbkpn replicaset-controller logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:21 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-598dc889b8 to 1 deployment-controller logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:22 +0000 UTC Normal Pod my-jaeger-collector-598dc889b8-qbkpn AddedInterface Add eth0 [10.128.2.56/23] from ovn-kubernetes multus logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:22 +0000 UTC Normal Pod my-jaeger-collector-598dc889b8-qbkpn.spec.containers{jaeger-collector} Pulling Pulling image "test" kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:22 +0000 UTC Warning Pod my-jaeger-collector-598dc889b8-qbkpn.spec.containers{jaeger-collector} Failed Failed to pull image "test": initializing source docker://test:latest: reading manifest latest in docker.io/library/test: requested access to the resource is denied kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:22 +0000 UTC Warning Pod my-jaeger-collector-598dc889b8-qbkpn.spec.containers{jaeger-collector} Failed Error: ErrImagePull kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:22 +0000 UTC Normal Pod my-jaeger-collector-598dc889b8-qbkpn.spec.containers{jaeger-collector} BackOff Back-off pulling image "test" kubelet logger.go:42: 20:38:23 | set-custom-img | 2025-02-17 20:38:22 +0000 UTC Warning Pod my-jaeger-collector-598dc889b8-qbkpn.spec.containers{jaeger-collector} Failed Error: ImagePullBackOff kubelet logger.go:42: 20:38:23 | set-custom-img | Deleting namespace: kuttl-test-perfect-monitor === CONT kuttl/harness/collector-otlp-production-http logger.go:42: 20:38:30 | collector-otlp-production-http | Creating namespace: kuttl-test-upright-primate logger.go:42: 20:38:30 | collector-otlp-production-http/1-install | starting test step 1-install logger.go:42: 20:38:30 | collector-otlp-production-http/1-install | Jaeger:kuttl-test-upright-primate/my-jaeger created logger.go:42: 20:39:07 | collector-otlp-production-http/1-install | test step completed 1-install logger.go:42: 20:39:07 | collector-otlp-production-http/2-smoke-test | starting test step 2-smoke-test logger.go:42: 20:39:07 | collector-otlp-production-http/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 20:39:08 | collector-otlp-production-http/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:39:14 | collector-otlp-production-http/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 20:39:15 | collector-otlp-production-http/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 20:39:15 | collector-otlp-production-http/2-smoke-test | job.batch/report-span created logger.go:42: 20:39:15 | collector-otlp-production-http/2-smoke-test | job.batch/check-span created logger.go:42: 20:39:27 | collector-otlp-production-http/2-smoke-test | test step completed 2-smoke-test logger.go:42: 20:39:27 | collector-otlp-production-http | collector-otlp-production-http events from ns kuttl-test-upright-primate: logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:38:37 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestuprightprimatemyjaeger-1-546477fcf8 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestuprightprimatemyjaeger-1-546477qnhwb replicaset-controller logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:38:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuprightprimatemyjaeger-1-546477qnhwb Binding Scheduled Successfully assigned kuttl-test-upright-primate/elasticsearch-cdm-kuttltestuprightprimatemyjaeger-1-546477qnhwb to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:38:37 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestuprightprimatemyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestuprightprimatemyjaeger-1-546477fcf8 to 1 deployment-controller logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:38:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuprightprimatemyjaeger-1-546477qnhwb AddedInterface Add eth0 [10.129.2.54/23] from ovn-kubernetes multus logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:38:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuprightprimatemyjaeger-1-546477qnhwb.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:5117d0aee260f74d6ef77c7c71ed073b1ad45b456f58894539b5697bec878310" already present on machine kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:38:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuprightprimatemyjaeger-1-546477qnhwb.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:38:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuprightprimatemyjaeger-1-546477qnhwb.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:38:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuprightprimatemyjaeger-1-546477qnhwb.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d63d612cdca72f64c45960a9b30bcfd0c746daabc17407df92c62694d5bc4bb6" already present on machine kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:38:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuprightprimatemyjaeger-1-546477qnhwb.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:38:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestuprightprimatemyjaeger-1-546477qnhwb.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:38:48 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestuprightprimatemyjaeger-1-546477qnhwb.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:38:53 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestuprightprimatemyjaeger-1-546477qnhwb.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:04 +0000 UTC Normal Pod my-jaeger-collector-5758ff7b6f-vllzk Binding Scheduled Successfully assigned kuttl-test-upright-primate/my-jaeger-collector-5758ff7b6f-vllzk to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:04 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5758ff7b6f SuccessfulCreate Created pod: my-jaeger-collector-5758ff7b6f-vllzk replicaset-controller logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:04 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-5758ff7b6f to 1 deployment-controller logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:04 +0000 UTC Normal Pod my-jaeger-query-7bff64757-mj2n4 Binding Scheduled Successfully assigned kuttl-test-upright-primate/my-jaeger-query-7bff64757-mj2n4 to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:04 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7bff64757 SuccessfulCreate Created pod: my-jaeger-query-7bff64757-mj2n4 replicaset-controller logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:04 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7bff64757 to 1 deployment-controller logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:05 +0000 UTC Normal Pod my-jaeger-collector-5758ff7b6f-vllzk AddedInterface Add eth0 [10.128.2.57/23] from ovn-kubernetes multus logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:05 +0000 UTC Normal Pod my-jaeger-collector-5758ff7b6f-vllzk.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:05 +0000 UTC Normal Pod my-jaeger-collector-5758ff7b6f-vllzk.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:05 +0000 UTC Normal Pod my-jaeger-collector-5758ff7b6f-vllzk.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:05 +0000 UTC Normal Pod my-jaeger-query-7bff64757-mj2n4 AddedInterface Add eth0 [10.131.0.61/23] from ovn-kubernetes multus logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:05 +0000 UTC Normal Pod my-jaeger-query-7bff64757-mj2n4.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:05 +0000 UTC Normal Pod my-jaeger-query-7bff64757-mj2n4.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:05 +0000 UTC Normal Pod my-jaeger-query-7bff64757-mj2n4.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:05 +0000 UTC Normal Pod my-jaeger-query-7bff64757-mj2n4.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:05 +0000 UTC Normal Pod my-jaeger-query-7bff64757-mj2n4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:05 +0000 UTC Normal Pod my-jaeger-query-7bff64757-mj2n4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:05 +0000 UTC Normal Pod my-jaeger-query-7bff64757-mj2n4.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:05 +0000 UTC Normal Pod my-jaeger-query-7bff64757-mj2n4.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:05 +0000 UTC Normal Pod my-jaeger-query-7bff64757-mj2n4.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:11 +0000 UTC Normal Pod my-jaeger-query-7bff64757-mj2n4.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:11 +0000 UTC Normal Pod my-jaeger-query-7bff64757-mj2n4.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:11 +0000 UTC Normal Pod my-jaeger-query-7bff64757-mj2n4.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:11 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7bff64757 SuccessfulDelete Deleted pod: my-jaeger-query-7bff64757-mj2n4 replicaset-controller logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:11 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-7bff64757 to 0 from 1 deployment-controller logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:12 +0000 UTC Normal Pod my-jaeger-query-6999ddfb79-krk6h Binding Scheduled Successfully assigned kuttl-test-upright-primate/my-jaeger-query-6999ddfb79-krk6h to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:12 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-6999ddfb79 SuccessfulCreate Created pod: my-jaeger-query-6999ddfb79-krk6h replicaset-controller logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:12 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-6999ddfb79 to 1 deployment-controller logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:13 +0000 UTC Normal Pod my-jaeger-query-6999ddfb79-krk6h AddedInterface Add eth0 [10.131.0.62/23] from ovn-kubernetes multus logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:13 +0000 UTC Normal Pod my-jaeger-query-6999ddfb79-krk6h.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:13 +0000 UTC Normal Pod my-jaeger-query-6999ddfb79-krk6h.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:13 +0000 UTC Normal Pod my-jaeger-query-6999ddfb79-krk6h.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:13 +0000 UTC Normal Pod my-jaeger-query-6999ddfb79-krk6h.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:13 +0000 UTC Normal Pod my-jaeger-query-6999ddfb79-krk6h.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:13 +0000 UTC Normal Pod my-jaeger-query-6999ddfb79-krk6h.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:13 +0000 UTC Normal Pod my-jaeger-query-6999ddfb79-krk6h.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:13 +0000 UTC Normal Pod my-jaeger-query-6999ddfb79-krk6h.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:13 +0000 UTC Normal Pod my-jaeger-query-6999ddfb79-krk6h.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:15 +0000 UTC Normal Pod check-span-t9zkz Binding Scheduled Successfully assigned kuttl-test-upright-primate/check-span-t9zkz to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:15 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-t9zkz job-controller logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:15 +0000 UTC Normal Pod report-span-49hd9 Binding Scheduled Successfully assigned kuttl-test-upright-primate/report-span-49hd9 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:15 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-49hd9 job-controller logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:16 +0000 UTC Normal Pod check-span-t9zkz AddedInterface Add eth0 [10.128.2.59/23] from ovn-kubernetes multus logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:16 +0000 UTC Normal Pod check-span-t9zkz.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:16 +0000 UTC Normal Pod check-span-t9zkz.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 555ms (555ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:16 +0000 UTC Normal Pod check-span-t9zkz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:16 +0000 UTC Normal Pod check-span-t9zkz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:16 +0000 UTC Normal Pod report-span-49hd9 AddedInterface Add eth0 [10.128.2.58/23] from ovn-kubernetes multus logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:16 +0000 UTC Normal Pod report-span-49hd9.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:16 +0000 UTC Normal Pod report-span-49hd9.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 583ms (583ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:16 +0000 UTC Normal Pod report-span-49hd9.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:16 +0000 UTC Normal Pod report-span-49hd9.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:39:27 | collector-otlp-production-http | 2025-02-17 20:39:27 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 20:39:27 | collector-otlp-production-http | Deleting namespace: kuttl-test-upright-primate === CONT kuttl/harness/collector-otlp-production-grpc logger.go:42: 20:39:41 | collector-otlp-production-grpc | Creating namespace: kuttl-test-inviting-worm logger.go:42: 20:39:41 | collector-otlp-production-grpc/1-install | starting test step 1-install logger.go:42: 20:39:41 | collector-otlp-production-grpc/1-install | Jaeger:kuttl-test-inviting-worm/my-jaeger created logger.go:42: 20:40:20 | collector-otlp-production-grpc/1-install | test step completed 1-install logger.go:42: 20:40:20 | collector-otlp-production-grpc/2-smoke-test | starting test step 2-smoke-test logger.go:42: 20:40:20 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 20:40:22 | collector-otlp-production-grpc/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:40:28 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 20:40:28 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 20:40:29 | collector-otlp-production-grpc/2-smoke-test | job.batch/report-span created logger.go:42: 20:40:29 | collector-otlp-production-grpc/2-smoke-test | job.batch/check-span created logger.go:42: 20:40:49 | collector-otlp-production-grpc/2-smoke-test | test step completed 2-smoke-test logger.go:42: 20:40:49 | collector-otlp-production-grpc | collector-otlp-production-grpc events from ns kuttl-test-inviting-worm: logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:39:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinvitingwormmyjaeger-1-5cc47f5c4bb6m Binding Scheduled Successfully assigned kuttl-test-inviting-worm/elasticsearch-cdm-kuttltestinvitingwormmyjaeger-1-5cc47f5c4bb6m to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:39:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinvitingwormmyjaeger-1-5cc47f5c4bb6m AddedInterface Add eth0 [10.129.2.55/23] from ovn-kubernetes multus logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:39:50 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestinvitingwormmyjaeger-1-5cc47f5c9d SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestinvitingwormmyjaeger-1-5cc47f5c4bb6m replicaset-controller logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:39:50 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestinvitingwormmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestinvitingwormmyjaeger-1-5cc47f5c9d to 1 deployment-controller logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:39:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinvitingwormmyjaeger-1-5cc47f5c4bb6m.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:5117d0aee260f74d6ef77c7c71ed073b1ad45b456f58894539b5697bec878310" already present on machine kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:39:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinvitingwormmyjaeger-1-5cc47f5c4bb6m.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:39:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinvitingwormmyjaeger-1-5cc47f5c4bb6m.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:39:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinvitingwormmyjaeger-1-5cc47f5c4bb6m.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d63d612cdca72f64c45960a9b30bcfd0c746daabc17407df92c62694d5bc4bb6" already present on machine kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:39:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinvitingwormmyjaeger-1-5cc47f5c4bb6m.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:39:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinvitingwormmyjaeger-1-5cc47f5c4bb6m.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:05 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestinvitingwormmyjaeger-1-5cc47f5c4bb6m.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:17 +0000 UTC Normal Pod my-jaeger-collector-64fc6c4bc9-wf6zw Binding Scheduled Successfully assigned kuttl-test-inviting-worm/my-jaeger-collector-64fc6c4bc9-wf6zw to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:17 +0000 UTC Normal Pod my-jaeger-collector-64fc6c4bc9-wf6zw AddedInterface Add eth0 [10.131.0.63/23] from ovn-kubernetes multus logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:17 +0000 UTC Normal Pod my-jaeger-collector-64fc6c4bc9-wf6zw.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:17 +0000 UTC Normal Pod my-jaeger-collector-64fc6c4bc9-wf6zw.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:17 +0000 UTC Normal Pod my-jaeger-collector-64fc6c4bc9-wf6zw.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:17 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-64fc6c4bc9 SuccessfulCreate Created pod: my-jaeger-collector-64fc6c4bc9-wf6zw replicaset-controller logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:17 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-64fc6c4bc9 to 1 deployment-controller logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:17 +0000 UTC Normal Pod my-jaeger-query-65f6c9b78d-tg9qd Binding Scheduled Successfully assigned kuttl-test-inviting-worm/my-jaeger-query-65f6c9b78d-tg9qd to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:17 +0000 UTC Normal Pod my-jaeger-query-65f6c9b78d-tg9qd AddedInterface Add eth0 [10.128.2.60/23] from ovn-kubernetes multus logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:17 +0000 UTC Normal Pod my-jaeger-query-65f6c9b78d-tg9qd.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:17 +0000 UTC Normal Pod my-jaeger-query-65f6c9b78d-tg9qd.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:17 +0000 UTC Normal Pod my-jaeger-query-65f6c9b78d-tg9qd.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:17 +0000 UTC Normal Pod my-jaeger-query-65f6c9b78d-tg9qd.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:17 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-65f6c9b78d SuccessfulCreate Created pod: my-jaeger-query-65f6c9b78d-tg9qd replicaset-controller logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:17 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-65f6c9b78d to 1 deployment-controller logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:18 +0000 UTC Normal Pod my-jaeger-query-65f6c9b78d-tg9qd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:18 +0000 UTC Normal Pod my-jaeger-query-65f6c9b78d-tg9qd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:18 +0000 UTC Normal Pod my-jaeger-query-65f6c9b78d-tg9qd.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:18 +0000 UTC Normal Pod my-jaeger-query-65f6c9b78d-tg9qd.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:18 +0000 UTC Normal Pod my-jaeger-query-65f6c9b78d-tg9qd.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:25 +0000 UTC Normal Pod my-jaeger-query-65f6c9b78d-tg9qd.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:25 +0000 UTC Normal Pod my-jaeger-query-65f6c9b78d-tg9qd.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:25 +0000 UTC Normal Pod my-jaeger-query-65f6c9b78d-tg9qd.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:25 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-65f6c9b78d SuccessfulDelete Deleted pod: my-jaeger-query-65f6c9b78d-tg9qd replicaset-controller logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:25 +0000 UTC Normal Pod my-jaeger-query-68df67864d-q2dll Binding Scheduled Successfully assigned kuttl-test-inviting-worm/my-jaeger-query-68df67864d-q2dll to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:25 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-68df67864d SuccessfulCreate Created pod: my-jaeger-query-68df67864d-q2dll replicaset-controller logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:25 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-65f6c9b78d to 0 from 1 deployment-controller logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:25 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-68df67864d to 1 deployment-controller logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:26 +0000 UTC Normal Pod my-jaeger-query-68df67864d-q2dll AddedInterface Add eth0 [10.128.2.61/23] from ovn-kubernetes multus logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:26 +0000 UTC Normal Pod my-jaeger-query-68df67864d-q2dll.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:26 +0000 UTC Normal Pod my-jaeger-query-68df67864d-q2dll.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:26 +0000 UTC Normal Pod my-jaeger-query-68df67864d-q2dll.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:26 +0000 UTC Normal Pod my-jaeger-query-68df67864d-q2dll.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:26 +0000 UTC Normal Pod my-jaeger-query-68df67864d-q2dll.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:26 +0000 UTC Normal Pod my-jaeger-query-68df67864d-q2dll.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:26 +0000 UTC Normal Pod my-jaeger-query-68df67864d-q2dll.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:26 +0000 UTC Normal Pod my-jaeger-query-68df67864d-q2dll.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:26 +0000 UTC Normal Pod my-jaeger-query-68df67864d-q2dll.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:29 +0000 UTC Normal Pod check-span-wdjfd Binding Scheduled Successfully assigned kuttl-test-inviting-worm/check-span-wdjfd to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:29 +0000 UTC Normal Pod check-span-wdjfd AddedInterface Add eth0 [10.131.0.65/23] from ovn-kubernetes multus logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:29 +0000 UTC Normal Pod check-span-wdjfd.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:29 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-wdjfd job-controller logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:29 +0000 UTC Normal Pod report-span-mjjxv Binding Scheduled Successfully assigned kuttl-test-inviting-worm/report-span-mjjxv to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:29 +0000 UTC Normal Pod report-span-mjjxv AddedInterface Add eth0 [10.131.0.64/23] from ovn-kubernetes multus logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:29 +0000 UTC Normal Pod report-span-mjjxv.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:29 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-mjjxv job-controller logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:30 +0000 UTC Normal Pod check-span-wdjfd.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 549ms (549ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:30 +0000 UTC Normal Pod check-span-wdjfd.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:30 +0000 UTC Normal Pod check-span-wdjfd.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:30 +0000 UTC Normal Pod report-span-mjjxv.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 505ms (505ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:30 +0000 UTC Normal Pod report-span-mjjxv.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:30 +0000 UTC Normal Pod report-span-mjjxv.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:32 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:32 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:32 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:47 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:47 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-64fc6c4bc9-wf6zw horizontal-pod-autoscaler logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:47 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 20:40:49 | collector-otlp-production-grpc | 2025-02-17 20:40:48 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 20:40:49 | collector-otlp-production-grpc | Deleting namespace: kuttl-test-inviting-worm === CONT kuttl/harness/collector-autoscale logger.go:42: 20:41:02 | collector-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:41:02 | collector-autoscale | Creating namespace: kuttl-test-teaching-toucan logger.go:42: 20:41:02 | collector-autoscale/1-install | starting test step 1-install logger.go:42: 20:41:02 | collector-autoscale/1-install | Jaeger:kuttl-test-teaching-toucan/simple-prod created logger.go:42: 20:41:39 | collector-autoscale/1-install | test step completed 1-install logger.go:42: 20:41:39 | collector-autoscale/2- | starting test step 2- logger.go:42: 20:41:39 | collector-autoscale/2- | test step completed 2- logger.go:42: 20:41:39 | collector-autoscale | collector-autoscale events from ns kuttl-test-teaching-toucan: logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:09 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingtoucansimpleprod-1-859c57hdd Binding Scheduled Successfully assigned kuttl-test-teaching-toucan/elasticsearch-cdm-kuttltestteachingtoucansimpleprod-1-859c57hdd to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:09 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestteachingtoucansimpleprod-1-859ccbf8c6 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestteachingtoucansimpleprod-1-859c57hdd replicaset-controller logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:09 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestteachingtoucansimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestteachingtoucansimpleprod-1-859ccbf8c6 to 1 deployment-controller logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingtoucansimpleprod-1-859c57hdd AddedInterface Add eth0 [10.129.2.56/23] from ovn-kubernetes multus logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingtoucansimpleprod-1-859c57hdd.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:5117d0aee260f74d6ef77c7c71ed073b1ad45b456f58894539b5697bec878310" already present on machine kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingtoucansimpleprod-1-859c57hdd.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingtoucansimpleprod-1-859c57hdd.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingtoucansimpleprod-1-859c57hdd.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d63d612cdca72f64c45960a9b30bcfd0c746daabc17407df92c62694d5bc4bb6" already present on machine kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingtoucansimpleprod-1-859c57hdd.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingtoucansimpleprod-1-859c57hdd.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:25 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestteachingtoucansimpleprod-1-859c57hdd.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:36 +0000 UTC Normal Pod simple-prod-collector-cdd899ff-w56h5 Binding Scheduled Successfully assigned kuttl-test-teaching-toucan/simple-prod-collector-cdd899ff-w56h5 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:36 +0000 UTC Warning Pod simple-prod-collector-cdd899ff-w56h5 FailedMount MountVolume.SetUp failed for volume "simple-prod-collector-tls-config-volume" : secret "simple-prod-collector-headless-tls" not found kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:36 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-cdd899ff SuccessfulCreate Created pod: simple-prod-collector-cdd899ff-w56h5 replicaset-controller logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:36 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-cdd899ff to 1 deployment-controller logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:36 +0000 UTC Normal Pod simple-prod-query-7988dd9c98-tpxsj Binding Scheduled Successfully assigned kuttl-test-teaching-toucan/simple-prod-query-7988dd9c98-tpxsj to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:36 +0000 UTC Normal Pod simple-prod-query-7988dd9c98-tpxsj AddedInterface Add eth0 [10.131.0.66/23] from ovn-kubernetes multus logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:36 +0000 UTC Normal Pod simple-prod-query-7988dd9c98-tpxsj.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:36 +0000 UTC Normal ReplicaSet.apps simple-prod-query-7988dd9c98 SuccessfulCreate Created pod: simple-prod-query-7988dd9c98-tpxsj replicaset-controller logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:36 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-7988dd9c98 to 1 deployment-controller logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:37 +0000 UTC Normal Pod simple-prod-collector-cdd899ff-w56h5 AddedInterface Add eth0 [10.128.2.62/23] from ovn-kubernetes multus logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:37 +0000 UTC Normal Pod simple-prod-collector-cdd899ff-w56h5.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:37 +0000 UTC Normal Pod simple-prod-collector-cdd899ff-w56h5.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:37 +0000 UTC Normal Pod simple-prod-collector-cdd899ff-w56h5.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:37 +0000 UTC Normal Pod simple-prod-query-7988dd9c98-tpxsj.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:37 +0000 UTC Normal Pod simple-prod-query-7988dd9c98-tpxsj.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:37 +0000 UTC Normal Pod simple-prod-query-7988dd9c98-tpxsj.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:37 +0000 UTC Normal Pod simple-prod-query-7988dd9c98-tpxsj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:37 +0000 UTC Normal Pod simple-prod-query-7988dd9c98-tpxsj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:37 +0000 UTC Normal Pod simple-prod-query-7988dd9c98-tpxsj.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:37 +0000 UTC Normal Pod simple-prod-query-7988dd9c98-tpxsj.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:41:39 | collector-autoscale | 2025-02-17 20:41:37 +0000 UTC Normal Pod simple-prod-query-7988dd9c98-tpxsj.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:41:39 | collector-autoscale | Deleting namespace: kuttl-test-teaching-toucan === CONT kuttl/harness/collector-otlp-allinone-grpc logger.go:42: 20:41:46 | collector-otlp-allinone-grpc | Creating namespace: kuttl-test-star-colt logger.go:42: 20:41:46 | collector-otlp-allinone-grpc/0-install | starting test step 0-install logger.go:42: 20:41:46 | collector-otlp-allinone-grpc/0-install | Jaeger:kuttl-test-star-colt/my-jaeger created logger.go:42: 20:41:53 | collector-otlp-allinone-grpc/0-install | test step completed 0-install logger.go:42: 20:41:53 | collector-otlp-allinone-grpc/1-smoke-test | starting test step 1-smoke-test logger.go:42: 20:41:53 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 20:41:54 | collector-otlp-allinone-grpc/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:42:02 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 20:42:02 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 20:42:02 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/report-span created logger.go:42: 20:42:02 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/check-span created logger.go:42: 20:42:23 | collector-otlp-allinone-grpc/1-smoke-test | test step completed 1-smoke-test logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | collector-otlp-allinone-grpc events from ns kuttl-test-star-colt: logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:41:50 +0000 UTC Normal Pod my-jaeger-5f7f65c896-v87bz Binding Scheduled Successfully assigned kuttl-test-star-colt/my-jaeger-5f7f65c896-v87bz to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:41:50 +0000 UTC Normal ReplicaSet.apps my-jaeger-5f7f65c896 SuccessfulCreate Created pod: my-jaeger-5f7f65c896-v87bz replicaset-controller logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:41:50 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5f7f65c896 to 1 deployment-controller logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:41:51 +0000 UTC Normal Pod my-jaeger-5f7f65c896-v87bz AddedInterface Add eth0 [10.128.2.63/23] from ovn-kubernetes multus logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:41:51 +0000 UTC Normal Pod my-jaeger-5f7f65c896-v87bz.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:41:51 +0000 UTC Normal Pod my-jaeger-5f7f65c896-v87bz.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:41:51 +0000 UTC Normal Pod my-jaeger-5f7f65c896-v87bz.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:41:51 +0000 UTC Normal Pod my-jaeger-5f7f65c896-v87bz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:41:51 +0000 UTC Normal Pod my-jaeger-5f7f65c896-v87bz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:41:51 +0000 UTC Normal Pod my-jaeger-5f7f65c896-v87bz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:00 +0000 UTC Normal Pod my-jaeger-5f7f65c896-v87bz.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:00 +0000 UTC Normal Pod my-jaeger-5f7f65c896-v87bz.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:00 +0000 UTC Normal ReplicaSet.apps my-jaeger-5f7f65c896 SuccessfulDelete Deleted pod: my-jaeger-5f7f65c896-v87bz replicaset-controller logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:00 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-5f7f65c896 to 0 from 1 deployment-controller logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:01 +0000 UTC Normal Pod my-jaeger-d597ff7f6-46882 Binding Scheduled Successfully assigned kuttl-test-star-colt/my-jaeger-d597ff7f6-46882 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:01 +0000 UTC Normal Pod my-jaeger-d597ff7f6-46882 AddedInterface Add eth0 [10.128.2.64/23] from ovn-kubernetes multus logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:01 +0000 UTC Normal Pod my-jaeger-d597ff7f6-46882.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:01 +0000 UTC Normal Pod my-jaeger-d597ff7f6-46882.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:01 +0000 UTC Normal Pod my-jaeger-d597ff7f6-46882.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:01 +0000 UTC Normal Pod my-jaeger-d597ff7f6-46882.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:01 +0000 UTC Normal Pod my-jaeger-d597ff7f6-46882.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:01 +0000 UTC Normal Pod my-jaeger-d597ff7f6-46882.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:01 +0000 UTC Normal ReplicaSet.apps my-jaeger-d597ff7f6 SuccessfulCreate Created pod: my-jaeger-d597ff7f6-46882 replicaset-controller logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:01 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-d597ff7f6 to 1 deployment-controller logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:02 +0000 UTC Normal Pod check-span-2thft Binding Scheduled Successfully assigned kuttl-test-star-colt/check-span-2thft to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:02 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-2thft job-controller logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:02 +0000 UTC Normal Pod report-span-fbf7n Binding Scheduled Successfully assigned kuttl-test-star-colt/report-span-fbf7n to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:02 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-fbf7n job-controller logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:03 +0000 UTC Normal Pod check-span-2thft AddedInterface Add eth0 [10.131.0.67/23] from ovn-kubernetes multus logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:03 +0000 UTC Normal Pod check-span-2thft.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:03 +0000 UTC Normal Pod report-span-fbf7n AddedInterface Add eth0 [10.129.2.57/23] from ovn-kubernetes multus logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:03 +0000 UTC Normal Pod report-span-fbf7n.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:04 +0000 UTC Normal Pod check-span-2thft.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 597ms (597ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:04 +0000 UTC Normal Pod check-span-2thft.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:04 +0000 UTC Normal Pod check-span-2thft.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:04 +0000 UTC Normal Pod report-span-fbf7n.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 674ms (674ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:04 +0000 UTC Normal Pod report-span-fbf7n.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:04 +0000 UTC Normal Pod report-span-fbf7n.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | 2025-02-17 20:42:22 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 20:42:23 | collector-otlp-allinone-grpc | Deleting namespace: kuttl-test-star-colt === CONT kuttl/harness/cassandra-spark logger.go:42: 20:42:36 | cassandra-spark | Ignoring 01-assert.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:42:36 | cassandra-spark | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:42:36 | cassandra-spark | Creating namespace: kuttl-test-amusing-cod logger.go:42: 20:42:36 | cassandra-spark | cassandra-spark events from ns kuttl-test-amusing-cod: logger.go:42: 20:42:36 | cassandra-spark | Deleting namespace: kuttl-test-amusing-cod === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (348.13s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (7.05s) --- PASS: kuttl/harness/collector-otlp-allinone-http (39.81s) --- PASS: kuttl/harness/set-custom-img (49.56s) --- PASS: kuttl/harness/collector-otlp-production-http (70.63s) --- PASS: kuttl/harness/collector-otlp-production-grpc (80.90s) --- PASS: kuttl/harness/collector-autoscale (44.06s) --- PASS: kuttl/harness/collector-otlp-allinone-grpc (49.78s) --- PASS: kuttl/harness/cassandra-spark (6.15s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name miscellaneous --report --output /logs/artifacts/miscellaneous.xml ./artifacts/kuttl-report.xml time="2025-02-17T20:42:42Z" level=debug msg="Setting a new name for the test suites" time="2025-02-17T20:42:42Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-02-17T20:42:42Z" level=debug msg="normalizing test case names" time="2025-02-17T20:42:42Z" level=debug msg="miscellaneous/artifacts -> miscellaneous_artifacts" time="2025-02-17T20:42:42Z" level=debug msg="miscellaneous/collector-otlp-allinone-http -> miscellaneous_collector_otlp_allinone_http" time="2025-02-17T20:42:42Z" level=debug msg="miscellaneous/set-custom-img -> miscellaneous_set_custom_img" time="2025-02-17T20:42:42Z" level=debug msg="miscellaneous/collector-otlp-production-http -> miscellaneous_collector_otlp_production_http" time="2025-02-17T20:42:42Z" level=debug msg="miscellaneous/collector-otlp-production-grpc -> miscellaneous_collector_otlp_production_grpc" time="2025-02-17T20:42:42Z" level=debug msg="miscellaneous/collector-autoscale -> miscellaneous_collector_autoscale" time="2025-02-17T20:42:42Z" level=debug msg="miscellaneous/collector-otlp-allinone-grpc -> miscellaneous_collector_otlp_allinone_grpc" time="2025-02-17T20:42:42Z" level=debug msg="miscellaneous/cassandra-spark -> miscellaneous_cassandra_spark" +----------------------------------------------+--------+ | NAME | RESULT | +----------------------------------------------+--------+ | miscellaneous_artifacts | passed | | miscellaneous_collector_otlp_allinone_http | passed | | miscellaneous_set_custom_img | passed | | miscellaneous_collector_otlp_production_http | passed | | miscellaneous_collector_otlp_production_grpc | passed | | miscellaneous_collector_autoscale | passed | | miscellaneous_collector_otlp_allinone_grpc | passed | | miscellaneous_cassandra_spark | passed | +----------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh sidecar false true + '[' 3 -ne 3 ']' + test_suite_name=sidecar + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/sidecar.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-sidecar make[2]: Entering directory '/tmp/jaeger-tests' ./tests/e2e/sidecar/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-14-222249 True False 39m Cluster version is 4.18.0-0.nightly-2025-02-14-222249' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-14-222249 True False 39m Cluster version is 4.18.0-0.nightly-2025-02-14-222249' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/sidecar/render.sh ++ export SUITE_DIR=./tests/e2e/sidecar ++ SUITE_DIR=./tests/e2e/sidecar ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/sidecar ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + jaeger_service_name=order + start_test sidecar-deployment + '[' 1 -ne 1 ']' + test_name=sidecar-deployment + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-deployment' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-deployment\e[0m' Rendering files for test sidecar-deployment + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build + '[' _build '!=' _build ']' + mkdir -p sidecar-deployment + cd sidecar-deployment + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-namespace + '[' 1 -ne 1 ']' + test_name=sidecar-namespace + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-namespace' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-namespace\e[0m' Rendering files for test sidecar-namespace + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-deployment + '[' sidecar-deployment '!=' _build ']' + cd .. + mkdir -p sidecar-namespace + cd sidecar-namespace + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-skip-webhook + '[' 1 -ne 1 ']' + test_name=sidecar-skip-webhook + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-skip-webhook' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-skip-webhook\e[0m' Rendering files for test sidecar-skip-webhook + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-namespace + '[' sidecar-namespace '!=' _build ']' + cd .. + mkdir -p sidecar-skip-webhook + cd sidecar-skip-webhook + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running sidecar E2E tests' Running sidecar E2E tests + cd tests/e2e/sidecar/_build + set +e + KUBECONFIG=/tmp/kubeconfig-998185858 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/sidecar-deployment === PAUSE kuttl/harness/sidecar-deployment === RUN kuttl/harness/sidecar-namespace === PAUSE kuttl/harness/sidecar-namespace === RUN kuttl/harness/sidecar-skip-webhook === PAUSE kuttl/harness/sidecar-skip-webhook === CONT kuttl/harness/artifacts logger.go:42: 20:42:50 | artifacts | Creating namespace: kuttl-test-accurate-wasp logger.go:42: 20:42:50 | artifacts | artifacts events from ns kuttl-test-accurate-wasp: logger.go:42: 20:42:50 | artifacts | Deleting namespace: kuttl-test-accurate-wasp === CONT kuttl/harness/sidecar-namespace logger.go:42: 20:42:56 | sidecar-namespace | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:42:56 | sidecar-namespace | Creating namespace: kuttl-test-capable-yak logger.go:42: 20:42:56 | sidecar-namespace/0-install | starting test step 0-install logger.go:42: 20:42:56 | sidecar-namespace/0-install | Jaeger:kuttl-test-capable-yak/agent-as-sidecar created logger.go:42: 20:43:02 | sidecar-namespace/0-install | test step completed 0-install logger.go:42: 20:43:02 | sidecar-namespace/1-install | starting test step 1-install logger.go:42: 20:43:02 | sidecar-namespace/1-install | Deployment:kuttl-test-capable-yak/vertx-create-span-sidecar created logger.go:42: 20:43:03 | sidecar-namespace/1-install | test step completed 1-install logger.go:42: 20:43:03 | sidecar-namespace/2-enable-injection | starting test step 2-enable-injection logger.go:42: 20:43:03 | sidecar-namespace/2-enable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="true"] logger.go:42: 20:43:04 | sidecar-namespace/2-enable-injection | namespace/kuttl-test-capable-yak annotated logger.go:42: 20:43:06 | sidecar-namespace/2-enable-injection | test step completed 2-enable-injection logger.go:42: 20:43:06 | sidecar-namespace/3-find-service | starting test step 3-find-service logger.go:42: 20:43:06 | sidecar-namespace/3-find-service | Job:kuttl-test-capable-yak/00-find-service created logger.go:42: 20:43:19 | sidecar-namespace/3-find-service | test step completed 3-find-service logger.go:42: 20:43:19 | sidecar-namespace/4-other-instance | starting test step 4-other-instance logger.go:42: 20:43:19 | sidecar-namespace/4-other-instance | Jaeger:kuttl-test-capable-yak/agent-as-sidecar2 created logger.go:42: 20:43:31 | sidecar-namespace/4-other-instance | test step completed 4-other-instance logger.go:42: 20:43:31 | sidecar-namespace/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 20:43:31 | sidecar-namespace/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 20:43:31 | sidecar-namespace/6-find-service | starting test step 6-find-service logger.go:42: 20:43:32 | sidecar-namespace/6-find-service | Job:kuttl-test-capable-yak/01-find-service created logger.go:42: 20:43:52 | sidecar-namespace/6-find-service | test step completed 6-find-service logger.go:42: 20:43:52 | sidecar-namespace/7-disable-injection | starting test step 7-disable-injection logger.go:42: 20:43:52 | sidecar-namespace/7-disable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="false"] logger.go:42: 20:43:52 | sidecar-namespace/7-disable-injection | namespace/kuttl-test-capable-yak annotated logger.go:42: 20:43:54 | sidecar-namespace/7-disable-injection | test step completed 7-disable-injection logger.go:42: 20:43:54 | sidecar-namespace | sidecar-namespace events from ns kuttl-test-capable-yak: logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:00 +0000 UTC Normal Pod agent-as-sidecar-56c5c5c574-rpmf8 Binding Scheduled Successfully assigned kuttl-test-capable-yak/agent-as-sidecar-56c5c5c574-rpmf8 to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:00 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-56c5c5c574 SuccessfulCreate Created pod: agent-as-sidecar-56c5c5c574-rpmf8 replicaset-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:00 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-56c5c5c574 to 1 deployment-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:01 +0000 UTC Normal Pod agent-as-sidecar-56c5c5c574-rpmf8 AddedInterface Add eth0 [10.131.0.68/23] from ovn-kubernetes multus logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:01 +0000 UTC Normal Pod agent-as-sidecar-56c5c5c574-rpmf8.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:01 +0000 UTC Normal Pod agent-as-sidecar-56c5c5c574-rpmf8.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:01 +0000 UTC Normal Pod agent-as-sidecar-56c5c5c574-rpmf8.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:02 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-5q6s7 Binding Scheduled Successfully assigned kuttl-test-capable-yak/vertx-create-span-sidecar-797645c8fc-5q6s7 to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:02 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulCreate Created pod: vertx-create-span-sidecar-797645c8fc-5q6s7 replicaset-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:02 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-797645c8fc to 1 deployment-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:03 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-5q6s7 AddedInterface Add eth0 [10.129.2.58/23] from ovn-kubernetes multus logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:03 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-5q6s7.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:03 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-5q6s7.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:03 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-5q6s7.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:04 +0000 UTC Normal Pod vertx-create-span-sidecar-6954f847c-k9rjt Binding Scheduled Successfully assigned kuttl-test-capable-yak/vertx-create-span-sidecar-6954f847c-k9rjt to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:04 +0000 UTC Normal Pod vertx-create-span-sidecar-6954f847c-k9rjt AddedInterface Add eth0 [10.128.2.65/23] from ovn-kubernetes multus logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:04 +0000 UTC Normal Pod vertx-create-span-sidecar-6954f847c-k9rjt.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:04 +0000 UTC Normal Pod vertx-create-span-sidecar-6954f847c-k9rjt.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:04 +0000 UTC Normal Pod vertx-create-span-sidecar-6954f847c-k9rjt.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:04 +0000 UTC Normal Pod vertx-create-span-sidecar-6954f847c-k9rjt.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:04 +0000 UTC Normal Pod vertx-create-span-sidecar-6954f847c-k9rjt.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:04 +0000 UTC Normal Pod vertx-create-span-sidecar-6954f847c-k9rjt.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:04 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6954f847c SuccessfulCreate Created pod: vertx-create-span-sidecar-6954f847c-k9rjt replicaset-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:04 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6954f847c to 1 deployment-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:06 +0000 UTC Normal Pod 00-find-service-th7xn Binding Scheduled Successfully assigned kuttl-test-capable-yak/00-find-service-th7xn to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:06 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-th7xn job-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:07 +0000 UTC Normal Pod 00-find-service-th7xn AddedInterface Add eth0 [10.129.2.59/23] from ovn-kubernetes multus logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:07 +0000 UTC Normal Pod 00-find-service-th7xn.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:08 +0000 UTC Normal Pod 00-find-service-th7xn.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 883ms (883ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:08 +0000 UTC Normal Pod 00-find-service-th7xn.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:08 +0000 UTC Normal Pod 00-find-service-th7xn.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:11 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-5q6s7.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.58:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:11 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-5q6s7.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.58:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:12 +0000 UTC Warning Pod vertx-create-span-sidecar-6954f847c-k9rjt.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.65:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:12 +0000 UTC Warning Pod vertx-create-span-sidecar-6954f847c-k9rjt.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.65:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:13 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-5q6s7.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:13 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-5q6s7.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.58:8080/": read tcp 10.129.2.2:48466->10.129.2.58:8080: read: connection reset by peer kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:13 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-5q6s7.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.58:8080/": dial tcp 10.129.2.58:8080: connect: connection refused kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:14 +0000 UTC Normal Pod vertx-create-span-sidecar-6954f847c-k9rjt.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:15 +0000 UTC Warning Pod vertx-create-span-sidecar-6954f847c-k9rjt.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.65:8080/": read tcp 10.128.2.2:53418->10.128.2.65:8080: read: connection reset by peer kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:15 +0000 UTC Warning Pod vertx-create-span-sidecar-6954f847c-k9rjt.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.65:8080/": dial tcp 10.128.2.65:8080: connect: connection refused kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:18 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:23 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-5q6s7.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.58:8080/": read tcp 10.129.2.2:58750->10.129.2.58:8080: read: connection reset by peer kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:26 +0000 UTC Normal Pod agent-as-sidecar2-5d9b4fbf94-rqcq9 Binding Scheduled Successfully assigned kuttl-test-capable-yak/agent-as-sidecar2-5d9b4fbf94-rqcq9 to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:26 +0000 UTC Normal Pod agent-as-sidecar2-5d9b4fbf94-rqcq9 AddedInterface Add eth0 [10.129.2.60/23] from ovn-kubernetes multus logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:26 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-5d9b4fbf94 SuccessfulCreate Created pod: agent-as-sidecar2-5d9b4fbf94-rqcq9 replicaset-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:26 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-5d9b4fbf94 to 1 deployment-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:27 +0000 UTC Normal Pod agent-as-sidecar2-5d9b4fbf94-rqcq9.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:29 +0000 UTC Normal Pod agent-as-sidecar2-5d9b4fbf94-rqcq9.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" in 2.647s (2.647s including waiting). Image size: 144304495 bytes. kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:29 +0000 UTC Normal Pod agent-as-sidecar2-5d9b4fbf94-rqcq9.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:29 +0000 UTC Normal Pod agent-as-sidecar2-5d9b4fbf94-rqcq9.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:32 +0000 UTC Normal Pod 01-find-service-ds7kk Binding Scheduled Successfully assigned kuttl-test-capable-yak/01-find-service-ds7kk to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:32 +0000 UTC Normal Pod 01-find-service-ds7kk AddedInterface Add eth0 [10.129.2.61/23] from ovn-kubernetes multus logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:32 +0000 UTC Normal Pod 01-find-service-ds7kk.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:32 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-ds7kk job-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:32 +0000 UTC Normal Pod agent-as-sidecar-56c5c5c574-rpmf8.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:33 +0000 UTC Normal Pod 01-find-service-ds7kk.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 633ms (633ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:33 +0000 UTC Normal Pod 01-find-service-ds7kk.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:33 +0000 UTC Normal Pod 01-find-service-ds7kk.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:35 +0000 UTC Normal Pod vertx-create-span-sidecar-547cd68875-8jwcg Binding Scheduled Successfully assigned kuttl-test-capable-yak/vertx-create-span-sidecar-547cd68875-8jwcg to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:35 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-547cd68875 SuccessfulCreate Created pod: vertx-create-span-sidecar-547cd68875-8jwcg replicaset-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:35 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulDelete Deleted pod: vertx-create-span-sidecar-797645c8fc-5q6s7 replicaset-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:35 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-797645c8fc to 0 from 1 deployment-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:35 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-547cd68875 to 1 from 0 deployment-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:36 +0000 UTC Normal Pod vertx-create-span-sidecar-547cd68875-8jwcg AddedInterface Add eth0 [10.131.0.69/23] from ovn-kubernetes multus logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:36 +0000 UTC Normal Pod vertx-create-span-sidecar-547cd68875-8jwcg.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:40 +0000 UTC Normal Pod vertx-create-span-sidecar-547cd68875-8jwcg.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.908s (3.908s including waiting). Image size: 282912835 bytes. kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:40 +0000 UTC Normal Pod vertx-create-span-sidecar-547cd68875-8jwcg.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:40 +0000 UTC Normal Pod vertx-create-span-sidecar-547cd68875-8jwcg.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:40 +0000 UTC Normal Pod vertx-create-span-sidecar-547cd68875-8jwcg.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:40 +0000 UTC Normal Pod vertx-create-span-sidecar-547cd68875-8jwcg.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:40 +0000 UTC Normal Pod vertx-create-span-sidecar-547cd68875-8jwcg.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:48 +0000 UTC Warning Pod vertx-create-span-sidecar-547cd68875-8jwcg.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.69:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:48 +0000 UTC Warning Pod vertx-create-span-sidecar-547cd68875-8jwcg.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.69:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:51 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:51 +0000 UTC Normal Pod vertx-create-span-sidecar-547cd68875-8jwcg.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:51 +0000 UTC Warning Pod vertx-create-span-sidecar-547cd68875-8jwcg.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.69:8080/": read tcp 10.131.0.2:60988->10.131.0.69:8080: read: connection reset by peer kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:51 +0000 UTC Warning Pod vertx-create-span-sidecar-547cd68875-8jwcg.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.69:8080/": dial tcp 10.131.0.69:8080: connect: connection refused kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:51 +0000 UTC Normal Pod vertx-create-span-sidecar-547cd68875-8jwcg.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:52 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6954f847c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-6954f847c-k9rjt replicaset-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:52 +0000 UTC Normal Pod vertx-create-span-sidecar-6f87556f96-r5bpn Binding Scheduled Successfully assigned kuttl-test-capable-yak/vertx-create-span-sidecar-6f87556f96-r5bpn to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:52 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6f87556f96 SuccessfulCreate Created pod: vertx-create-span-sidecar-6f87556f96-r5bpn replicaset-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:52 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-6954f847c to 0 from 1 deployment-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:52 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6f87556f96 to 1 from 0 deployment-controller logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:53 +0000 UTC Normal Pod vertx-create-span-sidecar-6f87556f96-r5bpn AddedInterface Add eth0 [10.129.2.62/23] from ovn-kubernetes multus logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:53 +0000 UTC Normal Pod vertx-create-span-sidecar-6f87556f96-r5bpn.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:53 +0000 UTC Normal Pod vertx-create-span-sidecar-6f87556f96-r5bpn.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:43:54 | sidecar-namespace | 2025-02-17 20:43:53 +0000 UTC Normal Pod vertx-create-span-sidecar-6f87556f96-r5bpn.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:43:54 | sidecar-namespace | Deleting namespace: kuttl-test-capable-yak === CONT kuttl/harness/sidecar-skip-webhook logger.go:42: 20:44:02 | sidecar-skip-webhook | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:44:02 | sidecar-skip-webhook | Creating namespace: kuttl-test-up-tortoise logger.go:42: 20:44:02 | sidecar-skip-webhook/0-install | starting test step 0-install logger.go:42: 20:44:02 | sidecar-skip-webhook/0-install | Jaeger:kuttl-test-up-tortoise/agent-as-sidecar created logger.go:42: 20:44:09 | sidecar-skip-webhook/0-install | test step completed 0-install logger.go:42: 20:44:09 | sidecar-skip-webhook/1-install | starting test step 1-install logger.go:42: 20:44:09 | sidecar-skip-webhook/1-install | Deployment:kuttl-test-up-tortoise/vertx-create-span-sidecar created logger.go:42: 20:44:11 | sidecar-skip-webhook/1-install | test step completed 1-install logger.go:42: 20:44:11 | sidecar-skip-webhook/2-add-anotation-and-label | starting test step 2-add-anotation-and-label logger.go:42: 20:44:11 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name=jaeger-operator --namespace kuttl-test-up-tortoise] logger.go:42: 20:44:11 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar labeled logger.go:42: 20:44:11 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-up-tortoise] logger.go:42: 20:44:11 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 20:44:11 | sidecar-skip-webhook/2-add-anotation-and-label | test step completed 2-add-anotation-and-label logger.go:42: 20:44:11 | sidecar-skip-webhook/3-remove-label | starting test step 3-remove-label logger.go:42: 20:44:11 | sidecar-skip-webhook/3-remove-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name- --namespace kuttl-test-up-tortoise] logger.go:42: 20:44:11 | sidecar-skip-webhook/3-remove-label | deployment.apps/vertx-create-span-sidecar unlabeled logger.go:42: 20:44:14 | sidecar-skip-webhook/3-remove-label | test step completed 3-remove-label logger.go:42: 20:44:14 | sidecar-skip-webhook | sidecar-skip-webhook events from ns kuttl-test-up-tortoise: logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:06 +0000 UTC Normal Pod agent-as-sidecar-7b8685699f-gb7qx Binding Scheduled Successfully assigned kuttl-test-up-tortoise/agent-as-sidecar-7b8685699f-gb7qx to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:06 +0000 UTC Warning Pod agent-as-sidecar-7b8685699f-gb7qx FailedMount MountVolume.SetUp failed for volume "agent-as-sidecar-collector-tls-config-volume" : secret "agent-as-sidecar-collector-headless-tls" not found kubelet logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:06 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-7b8685699f SuccessfulCreate Created pod: agent-as-sidecar-7b8685699f-gb7qx replicaset-controller logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:06 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-7b8685699f to 1 deployment-controller logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:07 +0000 UTC Normal Pod agent-as-sidecar-7b8685699f-gb7qx AddedInterface Add eth0 [10.129.2.63/23] from ovn-kubernetes multus logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:07 +0000 UTC Normal Pod agent-as-sidecar-7b8685699f-gb7qx.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:07 +0000 UTC Normal Pod agent-as-sidecar-7b8685699f-gb7qx.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:07 +0000 UTC Normal Pod agent-as-sidecar-7b8685699f-gb7qx.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:09 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-4nfs5 Binding Scheduled Successfully assigned kuttl-test-up-tortoise/vertx-create-span-sidecar-797645c8fc-4nfs5 to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:09 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulCreate Created pod: vertx-create-span-sidecar-797645c8fc-4nfs5 replicaset-controller logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:09 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-797645c8fc to 1 deployment-controller logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:10 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-4nfs5 AddedInterface Add eth0 [10.129.2.64/23] from ovn-kubernetes multus logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:10 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-4nfs5.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:10 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-4nfs5.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:10 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-4nfs5.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:11 +0000 UTC Normal Pod vertx-create-span-sidecar-57b688d78d-zn25k Binding Scheduled Successfully assigned kuttl-test-up-tortoise/vertx-create-span-sidecar-57b688d78d-zn25k to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:11 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-57b688d78d SuccessfulCreate Created pod: vertx-create-span-sidecar-57b688d78d-zn25k replicaset-controller logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:11 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-57b688d78d to 1 deployment-controller logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:12 +0000 UTC Normal Pod vertx-create-span-sidecar-57b688d78d-zn25k AddedInterface Add eth0 [10.128.2.66/23] from ovn-kubernetes multus logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:12 +0000 UTC Normal Pod vertx-create-span-sidecar-57b688d78d-zn25k.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:12 +0000 UTC Normal Pod vertx-create-span-sidecar-57b688d78d-zn25k.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:12 +0000 UTC Normal Pod vertx-create-span-sidecar-57b688d78d-zn25k.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:12 +0000 UTC Normal Pod vertx-create-span-sidecar-57b688d78d-zn25k.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:12 +0000 UTC Normal Pod vertx-create-span-sidecar-57b688d78d-zn25k.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:44:14 | sidecar-skip-webhook | 2025-02-17 20:44:12 +0000 UTC Normal Pod vertx-create-span-sidecar-57b688d78d-zn25k.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:44:14 | sidecar-skip-webhook | Deleting namespace: kuttl-test-up-tortoise === CONT kuttl/harness/sidecar-deployment logger.go:42: 20:44:21 | sidecar-deployment | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:44:21 | sidecar-deployment | Creating namespace: kuttl-test-grand-jaguar logger.go:42: 20:44:21 | sidecar-deployment/0-install | starting test step 0-install logger.go:42: 20:44:21 | sidecar-deployment/0-install | Jaeger:kuttl-test-grand-jaguar/agent-as-sidecar created logger.go:42: 20:44:28 | sidecar-deployment/0-install | test step completed 0-install logger.go:42: 20:44:28 | sidecar-deployment/1-install | starting test step 1-install logger.go:42: 20:44:28 | sidecar-deployment/1-install | Deployment:kuttl-test-grand-jaguar/vertx-create-span-sidecar created logger.go:42: 20:44:30 | sidecar-deployment/1-install | test step completed 1-install logger.go:42: 20:44:30 | sidecar-deployment/2-enable-injection | starting test step 2-enable-injection logger.go:42: 20:44:30 | sidecar-deployment/2-enable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-grand-jaguar] logger.go:42: 20:44:30 | sidecar-deployment/2-enable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 20:44:32 | sidecar-deployment/2-enable-injection | test step completed 2-enable-injection logger.go:42: 20:44:32 | sidecar-deployment/3-find-service | starting test step 3-find-service logger.go:42: 20:44:33 | sidecar-deployment/3-find-service | Job:kuttl-test-grand-jaguar/00-find-service created logger.go:42: 20:44:45 | sidecar-deployment/3-find-service | test step completed 3-find-service logger.go:42: 20:44:45 | sidecar-deployment/4-other-instance | starting test step 4-other-instance logger.go:42: 20:44:45 | sidecar-deployment/4-other-instance | Jaeger:kuttl-test-grand-jaguar/agent-as-sidecar2 created logger.go:42: 20:44:51 | sidecar-deployment/4-other-instance | test step completed 4-other-instance logger.go:42: 20:44:51 | sidecar-deployment/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 20:44:51 | sidecar-deployment/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 20:44:51 | sidecar-deployment/6-find-service | starting test step 6-find-service logger.go:42: 20:44:51 | sidecar-deployment/6-find-service | Job:kuttl-test-grand-jaguar/01-find-service created logger.go:42: 20:45:12 | sidecar-deployment/6-find-service | test step completed 6-find-service logger.go:42: 20:45:12 | sidecar-deployment/7-disable-injection | starting test step 7-disable-injection logger.go:42: 20:45:12 | sidecar-deployment/7-disable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=false --namespace kuttl-test-grand-jaguar] logger.go:42: 20:45:13 | sidecar-deployment/7-disable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 20:45:15 | sidecar-deployment/7-disable-injection | test step completed 7-disable-injection logger.go:42: 20:45:15 | sidecar-deployment | sidecar-deployment events from ns kuttl-test-grand-jaguar: logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:25 +0000 UTC Normal Pod agent-as-sidecar-5767b54bd7-vs7pv Binding Scheduled Successfully assigned kuttl-test-grand-jaguar/agent-as-sidecar-5767b54bd7-vs7pv to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:25 +0000 UTC Warning Pod agent-as-sidecar-5767b54bd7-vs7pv FailedMount MountVolume.SetUp failed for volume "agent-as-sidecar-collector-tls-config-volume" : secret "agent-as-sidecar-collector-headless-tls" not found kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:25 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-5767b54bd7 SuccessfulCreate Created pod: agent-as-sidecar-5767b54bd7-vs7pv replicaset-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:25 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-5767b54bd7 to 1 deployment-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:26 +0000 UTC Normal Pod agent-as-sidecar-5767b54bd7-vs7pv AddedInterface Add eth0 [10.129.2.65/23] from ovn-kubernetes multus logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:26 +0000 UTC Normal Pod agent-as-sidecar-5767b54bd7-vs7pv.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:26 +0000 UTC Normal Pod agent-as-sidecar-5767b54bd7-vs7pv.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:26 +0000 UTC Normal Pod agent-as-sidecar-5767b54bd7-vs7pv.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:28 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-t7r4r Binding Scheduled Successfully assigned kuttl-test-grand-jaguar/vertx-create-span-sidecar-797645c8fc-t7r4r to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:28 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulCreate Created pod: vertx-create-span-sidecar-797645c8fc-t7r4r replicaset-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:28 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-797645c8fc to 1 deployment-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:29 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-t7r4r AddedInterface Add eth0 [10.129.2.66/23] from ovn-kubernetes multus logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:29 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-t7r4r.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:29 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-t7r4r.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:29 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-t7r4r.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:30 +0000 UTC Normal Pod vertx-create-span-sidecar-6d6c77f87b-7wslq Binding Scheduled Successfully assigned kuttl-test-grand-jaguar/vertx-create-span-sidecar-6d6c77f87b-7wslq to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:30 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6d6c77f87b SuccessfulCreate Created pod: vertx-create-span-sidecar-6d6c77f87b-7wslq replicaset-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:30 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6d6c77f87b to 1 deployment-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:31 +0000 UTC Normal Pod vertx-create-span-sidecar-6d6c77f87b-7wslq AddedInterface Add eth0 [10.128.2.67/23] from ovn-kubernetes multus logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:31 +0000 UTC Normal Pod vertx-create-span-sidecar-6d6c77f87b-7wslq.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:31 +0000 UTC Normal Pod vertx-create-span-sidecar-6d6c77f87b-7wslq.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:31 +0000 UTC Normal Pod vertx-create-span-sidecar-6d6c77f87b-7wslq.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:31 +0000 UTC Normal Pod vertx-create-span-sidecar-6d6c77f87b-7wslq.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:31 +0000 UTC Normal Pod vertx-create-span-sidecar-6d6c77f87b-7wslq.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:31 +0000 UTC Normal Pod vertx-create-span-sidecar-6d6c77f87b-7wslq.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:33 +0000 UTC Normal Pod 00-find-service-7kfr8 Binding Scheduled Successfully assigned kuttl-test-grand-jaguar/00-find-service-7kfr8 to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:33 +0000 UTC Normal Pod 00-find-service-7kfr8 AddedInterface Add eth0 [10.131.0.70/23] from ovn-kubernetes multus logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:33 +0000 UTC Normal Pod 00-find-service-7kfr8.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:33 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-7kfr8 job-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:34 +0000 UTC Normal Pod 00-find-service-7kfr8.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 819ms (819ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:34 +0000 UTC Normal Pod 00-find-service-7kfr8.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:34 +0000 UTC Normal Pod 00-find-service-7kfr8.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:37 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-t7r4r.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.66:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:37 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-t7r4r.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.66:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:39 +0000 UTC Warning Pod vertx-create-span-sidecar-6d6c77f87b-7wslq.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.67:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:39 +0000 UTC Warning Pod vertx-create-span-sidecar-6d6c77f87b-7wslq.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.67:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:39 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-t7r4r.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:39 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-t7r4r.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.66:8080/": read tcp 10.129.2.2:37508->10.129.2.66:8080: read: connection reset by peer kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:39 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-t7r4r.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.66:8080/": dial tcp 10.129.2.66:8080: connect: connection refused kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:41 +0000 UTC Normal Pod vertx-create-span-sidecar-6d6c77f87b-7wslq.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:41 +0000 UTC Warning Pod vertx-create-span-sidecar-6d6c77f87b-7wslq.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.67:8080/": read tcp 10.128.2.2:38412->10.128.2.67:8080: read: connection reset by peer kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:41 +0000 UTC Warning Pod vertx-create-span-sidecar-6d6c77f87b-7wslq.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.67:8080/": dial tcp 10.128.2.67:8080: connect: connection refused kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:44 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:49 +0000 UTC Normal Pod agent-as-sidecar2-748cb8bdd5-w6k8s Binding Scheduled Successfully assigned kuttl-test-grand-jaguar/agent-as-sidecar2-748cb8bdd5-w6k8s to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:49 +0000 UTC Normal Pod agent-as-sidecar2-748cb8bdd5-w6k8s AddedInterface Add eth0 [10.131.0.71/23] from ovn-kubernetes multus logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:49 +0000 UTC Normal Pod agent-as-sidecar2-748cb8bdd5-w6k8s.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:49 +0000 UTC Normal Pod agent-as-sidecar2-748cb8bdd5-w6k8s.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:49 +0000 UTC Normal Pod agent-as-sidecar2-748cb8bdd5-w6k8s.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:49 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-748cb8bdd5 SuccessfulCreate Created pod: agent-as-sidecar2-748cb8bdd5-w6k8s replicaset-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:49 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-748cb8bdd5 to 1 deployment-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:49 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-t7r4r.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.66:8080/": read tcp 10.129.2.2:56270->10.129.2.66:8080: read: connection reset by peer kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:51 +0000 UTC Normal Pod 01-find-service-5b6nr Binding Scheduled Successfully assigned kuttl-test-grand-jaguar/01-find-service-5b6nr to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:51 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-5b6nr job-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:51 +0000 UTC Normal Pod agent-as-sidecar-5767b54bd7-vs7pv.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:52 +0000 UTC Normal Pod 01-find-service-5b6nr AddedInterface Add eth0 [10.129.2.67/23] from ovn-kubernetes multus logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:52 +0000 UTC Normal Pod 01-find-service-5b6nr.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:53 +0000 UTC Normal Pod 01-find-service-5b6nr.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 918ms (918ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:53 +0000 UTC Normal Pod 01-find-service-5b6nr.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:53 +0000 UTC Normal Pod 01-find-service-5b6nr.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:55 +0000 UTC Normal Pod vertx-create-span-sidecar-777b98f9d6-wlbb8 Binding Scheduled Successfully assigned kuttl-test-grand-jaguar/vertx-create-span-sidecar-777b98f9d6-wlbb8 to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:55 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-777b98f9d6 SuccessfulCreate Created pod: vertx-create-span-sidecar-777b98f9d6-wlbb8 replicaset-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:55 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulDelete Deleted pod: vertx-create-span-sidecar-797645c8fc-t7r4r replicaset-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:55 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-797645c8fc to 0 from 1 deployment-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:55 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-777b98f9d6 to 1 from 0 deployment-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:56 +0000 UTC Normal Pod vertx-create-span-sidecar-777b98f9d6-wlbb8 AddedInterface Add eth0 [10.129.2.68/23] from ovn-kubernetes multus logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:56 +0000 UTC Normal Pod vertx-create-span-sidecar-777b98f9d6-wlbb8.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:56 +0000 UTC Normal Pod vertx-create-span-sidecar-777b98f9d6-wlbb8.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:56 +0000 UTC Normal Pod vertx-create-span-sidecar-777b98f9d6-wlbb8.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:56 +0000 UTC Normal Pod vertx-create-span-sidecar-777b98f9d6-wlbb8.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:58 +0000 UTC Normal Pod vertx-create-span-sidecar-777b98f9d6-wlbb8.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" in 1.999s (1.999s including waiting). Image size: 112614125 bytes. kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:58 +0000 UTC Normal Pod vertx-create-span-sidecar-777b98f9d6-wlbb8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:44:58 +0000 UTC Normal Pod vertx-create-span-sidecar-777b98f9d6-wlbb8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:45:04 +0000 UTC Warning Pod vertx-create-span-sidecar-777b98f9d6-wlbb8.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.68:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:45:04 +0000 UTC Warning Pod vertx-create-span-sidecar-777b98f9d6-wlbb8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.68:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:45:07 +0000 UTC Normal Pod vertx-create-span-sidecar-777b98f9d6-wlbb8.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:45:07 +0000 UTC Warning Pod vertx-create-span-sidecar-777b98f9d6-wlbb8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.68:8080/": read tcp 10.129.2.2:55990->10.129.2.68:8080: read: connection reset by peer kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:45:07 +0000 UTC Warning Pod vertx-create-span-sidecar-777b98f9d6-wlbb8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.68:8080/": dial tcp 10.129.2.68:8080: connect: connection refused kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:45:12 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:45:13 +0000 UTC Normal Pod vertx-create-span-sidecar-6757b5868f-7sjvh Binding Scheduled Successfully assigned kuttl-test-grand-jaguar/vertx-create-span-sidecar-6757b5868f-7sjvh to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:45:13 +0000 UTC Normal Pod vertx-create-span-sidecar-6757b5868f-7sjvh AddedInterface Add eth0 [10.131.0.72/23] from ovn-kubernetes multus logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:45:13 +0000 UTC Normal Pod vertx-create-span-sidecar-6757b5868f-7sjvh.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:45:13 +0000 UTC Normal Pod vertx-create-span-sidecar-6757b5868f-7sjvh.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:45:13 +0000 UTC Normal Pod vertx-create-span-sidecar-6757b5868f-7sjvh.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:45:13 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6757b5868f SuccessfulCreate Created pod: vertx-create-span-sidecar-6757b5868f-7sjvh replicaset-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:45:13 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6d6c77f87b SuccessfulDelete Deleted pod: vertx-create-span-sidecar-6d6c77f87b-7wslq replicaset-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:45:13 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-6d6c77f87b to 0 from 1 deployment-controller logger.go:42: 20:45:15 | sidecar-deployment | 2025-02-17 20:45:13 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6757b5868f to 1 from 0 deployment-controller logger.go:42: 20:45:15 | sidecar-deployment | Deleting namespace: kuttl-test-grand-jaguar === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (151.82s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.10s) --- PASS: kuttl/harness/sidecar-namespace (65.53s) --- PASS: kuttl/harness/sidecar-skip-webhook (19.14s) --- PASS: kuttl/harness/sidecar-deployment (60.91s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name sidecar --report --output /logs/artifacts/sidecar.xml ./artifacts/kuttl-report.xml time="2025-02-17T20:45:22Z" level=debug msg="Setting a new name for the test suites" time="2025-02-17T20:45:22Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-02-17T20:45:22Z" level=debug msg="normalizing test case names" time="2025-02-17T20:45:22Z" level=debug msg="sidecar/artifacts -> sidecar_artifacts" time="2025-02-17T20:45:22Z" level=debug msg="sidecar/sidecar-namespace -> sidecar_sidecar_namespace" time="2025-02-17T20:45:22Z" level=debug msg="sidecar/sidecar-skip-webhook -> sidecar_sidecar_skip_webhook" time="2025-02-17T20:45:22Z" level=debug msg="sidecar/sidecar-deployment -> sidecar_sidecar_deployment" +------------------------------+--------+ | NAME | RESULT | +------------------------------+--------+ | sidecar_artifacts | passed | | sidecar_sidecar_namespace | passed | | sidecar_sidecar_skip_webhook | passed | | sidecar_sidecar_deployment | passed | +------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh streaming false true + '[' 3 -ne 3 ']' + test_suite_name=streaming + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/streaming.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-streaming make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ SKIP_ES_EXTERNAL=true \ ./tests/e2e/streaming/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-14-222249 True False 41m Cluster version is 4.18.0-0.nightly-2025-02-14-222249' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-14-222249 True False 41m Cluster version is 4.18.0-0.nightly-2025-02-14-222249' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/streaming/render.sh ++ export SUITE_DIR=./tests/e2e/streaming ++ SUITE_DIR=./tests/e2e/streaming ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/streaming ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + '[' false = true ']' + start_test streaming-simple + '[' 1 -ne 1 ']' + test_name=streaming-simple + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-simple' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-simple\e[0m' Rendering files for test streaming-simple + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + mkdir -p streaming-simple + cd streaming-simple + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/streaming-jaeger-assert.yaml.template -o ./04-assert.yaml + render_smoke_test simple-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=simple-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + export JAEGER_NAME=simple-streaming + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-tls + '[' 1 -ne 1 ']' + test_name=streaming-with-tls + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-tls' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-tls\e[0m' Rendering files for test streaming-with-tls + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-simple + '[' streaming-simple '!=' _build ']' + cd .. + mkdir -p streaming-with-tls + cd streaming-with-tls + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + render_smoke_test tls-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=tls-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + export JAEGER_NAME=tls-streaming + JAEGER_NAME=tls-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-autoprovisioning-autoscale + '[' 1 -ne 1 ']' + test_name=streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-autoprovisioning-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-autoprovisioning-autoscale\e[0m' Rendering files for test streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-tls + '[' streaming-with-tls '!=' _build ']' + cd .. + mkdir -p streaming-with-autoprovisioning-autoscale + cd streaming-with-autoprovisioning-autoscale + '[' true = true ']' + rm ./00-install.yaml ./00-assert.yaml + render_install_elasticsearch upstream 01 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=01 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./01-assert.yaml + jaeger_name=auto-provisioned + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="20Mi"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="500m"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.autoscale=true ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.minReplicas=1 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.maxReplicas=2 ./02-install.yaml + render_assert_kafka true auto-provisioned 03 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provisioned + test_step=03 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./03-assert.yaml ++ expr 03 + 1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./04-assert.yaml ++ expr 03 + 2 + CLUSTER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./05-assert.yaml + version_lt 1.30 1.23 ++ echo 1.30 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.30 '!=' 1.30 + rm ./08-assert.yaml + skip_test streaming-with-tls 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-with-tls + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-autoprovisioning-autoscale + '[' streaming-with-autoprovisioning-autoscale '!=' _build ']' + cd .. + rm -rf streaming-with-tls + warning 'streaming-with-tls: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-with-tls: This test is flaky in Prow CI\e[0m' WAR: streaming-with-tls: This test is flaky in Prow CI + skip_test streaming-simple 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-simple + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + rm -rf streaming-simple + warning 'streaming-simple: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-simple: This test is flaky in Prow CI\e[0m' WAR: streaming-simple: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running streaming E2E tests' Running streaming E2E tests + cd tests/e2e/streaming/_build + set +e + KUBECONFIG=/tmp/kubeconfig-998185858 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 2 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/streaming-with-autoprovisioning-autoscale === PAUSE kuttl/harness/streaming-with-autoprovisioning-autoscale === CONT kuttl/harness/artifacts logger.go:42: 20:45:34 | artifacts | Creating namespace: kuttl-test-civil-chimp logger.go:42: 20:45:34 | artifacts | artifacts events from ns kuttl-test-civil-chimp: logger.go:42: 20:45:34 | artifacts | Deleting namespace: kuttl-test-civil-chimp === CONT kuttl/harness/streaming-with-autoprovisioning-autoscale logger.go:42: 20:45:41 | streaming-with-autoprovisioning-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:45:41 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:45:41 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:45:41 | streaming-with-autoprovisioning-autoscale | Creating namespace: kuttl-test-suited-fawn logger.go:42: 20:45:41 | streaming-with-autoprovisioning-autoscale/1-install | starting test step 1-install logger.go:42: 20:45:41 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc create sa deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 20:45:41 | streaming-with-autoprovisioning-autoscale/1-install | serviceaccount/deploy-elasticsearch created logger.go:42: 20:45:41 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc adm policy add-scc-to-user privileged -z deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 20:45:41 | streaming-with-autoprovisioning-autoscale/1-install | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:privileged added: "deploy-elasticsearch" logger.go:42: 20:45:41 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 6] logger.go:42: 20:45:47 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_0.yml -n $NAMESPACE] logger.go:42: 20:45:47 | streaming-with-autoprovisioning-autoscale/1-install | statefulset.apps/elasticsearch created logger.go:42: 20:45:47 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 3] logger.go:42: 20:45:50 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_1.yml -n $NAMESPACE] logger.go:42: 20:45:50 | streaming-with-autoprovisioning-autoscale/1-install | service/elasticsearch created logger.go:42: 20:46:08 | streaming-with-autoprovisioning-autoscale/1-install | test step completed 1-install logger.go:42: 20:46:08 | streaming-with-autoprovisioning-autoscale/2-install | starting test step 2-install logger.go:42: 20:46:08 | streaming-with-autoprovisioning-autoscale/2-install | Jaeger:kuttl-test-suited-fawn/auto-provisioned created logger.go:42: 20:46:08 | streaming-with-autoprovisioning-autoscale/2-install | test step completed 2-install logger.go:42: 20:46:08 | streaming-with-autoprovisioning-autoscale/3- | starting test step 3- logger.go:42: 20:46:53 | streaming-with-autoprovisioning-autoscale/3- | test step completed 3- logger.go:42: 20:46:53 | streaming-with-autoprovisioning-autoscale/4- | starting test step 4- logger.go:42: 20:47:24 | streaming-with-autoprovisioning-autoscale/4- | test step completed 4- logger.go:42: 20:47:24 | streaming-with-autoprovisioning-autoscale/5- | starting test step 5- logger.go:42: 20:47:47 | streaming-with-autoprovisioning-autoscale/5- | test step completed 5- logger.go:42: 20:47:47 | streaming-with-autoprovisioning-autoscale/6- | starting test step 6- logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale/6- | test step completed 6- logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale/7- | starting test step 7- logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale/7- | test step completed 7- logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | streaming-with-autoprovisioning-autoscale events from ns kuttl-test-suited-fawn: logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:45:47 +0000 UTC Normal Pod elasticsearch-0 Binding Scheduled Successfully assigned kuttl-test-suited-fawn/elasticsearch-0 to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:45:47 +0000 UTC Normal StatefulSet.apps elasticsearch SuccessfulCreate create Pod elasticsearch-0 in StatefulSet elasticsearch successful statefulset-controller logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:45:48 +0000 UTC Normal Pod elasticsearch-0 AddedInterface Add eth0 [10.129.2.70/23] from ovn-kubernetes multus logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:45:48 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulling Pulling image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:45:55 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulled Successfully pulled image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" in 6.79s (6.79s including waiting). Image size: 758467647 bytes. kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:45:55 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:45:55 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:02 +0000 UTC Warning Pod elasticsearch-0.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Get "http://10.129.2.70:9200/": dial tcp 10.129.2.70:9200: connect: connection refused kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:16 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-zookeeper NoPods No matching pods found controllermanager logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:16 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:16 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-suited-fawn/data-auto-provisioned-zookeeper-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-6bc94f46d8-4qw9c_3bae061c-6de9-4742-b3a9-fe5a89ec2255 logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:16 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:18 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ProvisioningSucceeded Successfully provisioned volume pvc-0803c03a-33a7-442a-840b-9603742844cb ebs.csi.aws.com_aws-ebs-csi-driver-controller-6bc94f46d8-4qw9c_3bae061c-6de9-4742-b3a9-fe5a89ec2255 logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:19 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 Binding Scheduled Successfully assigned kuttl-test-suited-fawn/auto-provisioned-zookeeper-0 to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:21 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-0803c03a-33a7-442a-840b-9603742844cb" attachdetach-controller logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:23 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 AddedInterface Add eth0 [10.131.0.73/23] from ovn-kubernetes multus logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:23 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:e771a84a458388c07ecf970ec625e4203c18e5d2bd91e7571d9c44ebb09313de" kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:32 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:e771a84a458388c07ecf970ec625e4203c18e5d2bd91e7571d9c44ebb09313de" in 9.246s (9.246s including waiting). Image size: 617327760 bytes. kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:32 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Created Created container zookeeper kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:32 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Started Started container zookeeper kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:54 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-kafka NoPods No matching pods found controllermanager logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:54 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:54 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-suited-fawn/data-0-auto-provisioned-kafka-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-6bc94f46d8-4qw9c_3bae061c-6de9-4742-b3a9-fe5a89ec2255 logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:54 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:56 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ProvisioningSucceeded Successfully provisioned volume pvc-1ab534f6-15e2-439b-9caa-2dcdd1187628 ebs.csi.aws.com_aws-ebs-csi-driver-controller-6bc94f46d8-4qw9c_3bae061c-6de9-4742-b3a9-fe5a89ec2255 logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:46:57 +0000 UTC Normal Pod auto-provisioned-kafka-0 Binding Scheduled Successfully assigned kuttl-test-suited-fawn/auto-provisioned-kafka-0 to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:03 +0000 UTC Normal Pod auto-provisioned-kafka-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-1ab534f6-15e2-439b-9caa-2dcdd1187628" attachdetach-controller logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:04 +0000 UTC Normal Pod auto-provisioned-kafka-0 AddedInterface Add eth0 [10.131.0.74/23] from ovn-kubernetes multus logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:04 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulled Container image "registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:e771a84a458388c07ecf970ec625e4203c18e5d2bd91e7571d9c44ebb09313de" already present on machine kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:04 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Created Created container kafka kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:04 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Started Started container kafka kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:25 +0000 UTC Normal Pod auto-provisioned-entity-operator-654fc444f9-x279m Binding Scheduled Successfully assigned kuttl-test-suited-fawn/auto-provisioned-entity-operator-654fc444f9-x279m to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:25 +0000 UTC Normal ReplicaSet.apps auto-provisioned-entity-operator-654fc444f9 SuccessfulCreate Created pod: auto-provisioned-entity-operator-654fc444f9-x279m replicaset-controller logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:25 +0000 UTC Normal Deployment.apps auto-provisioned-entity-operator ScalingReplicaSet Scaled up replica set auto-provisioned-entity-operator-654fc444f9 to 1 deployment-controller logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:26 +0000 UTC Normal Pod auto-provisioned-entity-operator-654fc444f9-x279m AddedInterface Add eth0 [10.129.2.71/23] from ovn-kubernetes multus logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:26 +0000 UTC Normal Pod auto-provisioned-entity-operator-654fc444f9-x279m.spec.containers{topic-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec" already present on machine kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:26 +0000 UTC Normal Pod auto-provisioned-entity-operator-654fc444f9-x279m.spec.containers{topic-operator} Created Created container topic-operator kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:26 +0000 UTC Normal Pod auto-provisioned-entity-operator-654fc444f9-x279m.spec.containers{topic-operator} Started Started container topic-operator kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:26 +0000 UTC Normal Pod auto-provisioned-entity-operator-654fc444f9-x279m.spec.containers{user-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:675d4acc375fb20e1a3fd5755184a318341c15f9337a11ea41e902fcda6d0bec" already present on machine kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:26 +0000 UTC Normal Pod auto-provisioned-entity-operator-654fc444f9-x279m.spec.containers{user-operator} Created Created container user-operator kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:26 +0000 UTC Normal Pod auto-provisioned-entity-operator-654fc444f9-x279m.spec.containers{user-operator} Started Started container user-operator kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:48 +0000 UTC Normal Pod auto-provisioned-collector-688c5b6b55-4kv96 Binding Scheduled Successfully assigned kuttl-test-suited-fawn/auto-provisioned-collector-688c5b6b55-4kv96 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:48 +0000 UTC Normal ReplicaSet.apps auto-provisioned-collector-688c5b6b55 SuccessfulCreate Created pod: auto-provisioned-collector-688c5b6b55-4kv96 replicaset-controller logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:48 +0000 UTC Normal Deployment.apps auto-provisioned-collector ScalingReplicaSet Scaled up replica set auto-provisioned-collector-688c5b6b55 to 1 deployment-controller logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:48 +0000 UTC Normal Pod auto-provisioned-ingester-cdc895fcc-q8wmb Binding Scheduled Successfully assigned kuttl-test-suited-fawn/auto-provisioned-ingester-cdc895fcc-q8wmb to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:48 +0000 UTC Normal ReplicaSet.apps auto-provisioned-ingester-cdc895fcc SuccessfulCreate Created pod: auto-provisioned-ingester-cdc895fcc-q8wmb replicaset-controller logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:48 +0000 UTC Normal Deployment.apps auto-provisioned-ingester ScalingReplicaSet Scaled up replica set auto-provisioned-ingester-cdc895fcc to 1 deployment-controller logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:48 +0000 UTC Normal Pod auto-provisioned-query-58d7859799-h75r8 Binding Scheduled Successfully assigned kuttl-test-suited-fawn/auto-provisioned-query-58d7859799-h75r8 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:48 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-58d7859799 SuccessfulCreate Created pod: auto-provisioned-query-58d7859799-h75r8 replicaset-controller logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:48 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled up replica set auto-provisioned-query-58d7859799 to 1 deployment-controller logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:49 +0000 UTC Normal Pod auto-provisioned-ingester-cdc895fcc-q8wmb AddedInterface Add eth0 [10.128.2.68/23] from ovn-kubernetes multus logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:49 +0000 UTC Normal Pod auto-provisioned-ingester-cdc895fcc-q8wmb.spec.containers{jaeger-ingester} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:b06a13d35b7ff03ed3a5512b64e5a472a8365e99dc5c9ba69e844db2b5284ee8" kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:49 +0000 UTC Normal Pod auto-provisioned-query-58d7859799-h75r8 AddedInterface Add eth0 [10.128.2.70/23] from ovn-kubernetes multus logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:49 +0000 UTC Normal Pod auto-provisioned-query-58d7859799-h75r8.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:49 +0000 UTC Normal Pod auto-provisioned-query-58d7859799-h75r8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:49 +0000 UTC Normal Pod auto-provisioned-query-58d7859799-h75r8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:49 +0000 UTC Normal Pod auto-provisioned-query-58d7859799-h75r8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:49 +0000 UTC Normal Pod auto-provisioned-query-58d7859799-h75r8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:49 +0000 UTC Normal Pod auto-provisioned-query-58d7859799-h75r8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:49 +0000 UTC Normal Pod auto-provisioned-query-58d7859799-h75r8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:49 +0000 UTC Normal Pod auto-provisioned-query-58d7859799-h75r8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:49 +0000 UTC Normal Pod auto-provisioned-query-58d7859799-h75r8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:50 +0000 UTC Normal Pod auto-provisioned-collector-688c5b6b55-4kv96 AddedInterface Add eth0 [10.128.2.69/23] from ovn-kubernetes multus logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:50 +0000 UTC Normal Pod auto-provisioned-collector-688c5b6b55-4kv96.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:50 +0000 UTC Normal Pod auto-provisioned-collector-688c5b6b55-4kv96.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:50 +0000 UTC Normal Pod auto-provisioned-collector-688c5b6b55-4kv96.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:53 +0000 UTC Normal Pod auto-provisioned-ingester-cdc895fcc-q8wmb.spec.containers{jaeger-ingester} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:b06a13d35b7ff03ed3a5512b64e5a472a8365e99dc5c9ba69e844db2b5284ee8" in 3.716s (3.716s including waiting). Image size: 137309437 bytes. kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:53 +0000 UTC Normal Pod auto-provisioned-ingester-cdc895fcc-q8wmb.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:53 +0000 UTC Normal Pod auto-provisioned-ingester-cdc895fcc-q8wmb.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | 2025-02-17 20:47:54 +0000 UTC Warning Pod auto-provisioned-ingester-cdc895fcc-q8wmb.spec.containers{jaeger-ingester} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 20:47:59 | streaming-with-autoprovisioning-autoscale | Deleting namespace: kuttl-test-suited-fawn === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (164.88s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.53s) --- PASS: kuttl/harness/streaming-with-autoprovisioning-autoscale (158.14s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name streaming --report --output /logs/artifacts/streaming.xml ./artifacts/kuttl-report.xml time="2025-02-17T20:48:19Z" level=debug msg="Setting a new name for the test suites" time="2025-02-17T20:48:19Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-02-17T20:48:19Z" level=debug msg="normalizing test case names" time="2025-02-17T20:48:19Z" level=debug msg="streaming/artifacts -> streaming_artifacts" time="2025-02-17T20:48:19Z" level=debug msg="streaming/streaming-with-autoprovisioning-autoscale -> streaming_streaming_with_autoprovisioning_autoscale" +-----------------------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------------------+--------+ | streaming_artifacts | passed | | streaming_streaming_with_autoprovisioning_autoscale | passed | +-----------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh ui false true + '[' 3 -ne 3 ']' + test_suite_name=ui + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/ui.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-ui make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true ./tests/e2e/ui/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-14-222249 True False 44m Cluster version is 4.18.0-0.nightly-2025-02-14-222249' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-14-222249 True False 44m Cluster version is 4.18.0-0.nightly-2025-02-14-222249' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/ui/render.sh ++ export SUITE_DIR=./tests/e2e/ui ++ SUITE_DIR=./tests/e2e/ui ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/ui ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test allinone + '[' 1 -ne 1 ']' + test_name=allinone + echo =========================================================================== =========================================================================== + info 'Rendering files for test allinone' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test allinone\e[0m' Rendering files for test allinone + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build + '[' _build '!=' _build ']' + mkdir -p allinone + cd allinone + export GET_URL_COMMAND + export URL + export JAEGER_NAME=all-in-one-ui + JAEGER_NAME=all-in-one-ui + '[' true = true ']' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./01-curl.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./04-test-ui-config.yaml + start_test production + '[' 1 -ne 1 ']' + test_name=production + echo =========================================================================== =========================================================================== + info 'Rendering files for test production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test production\e[0m' Rendering files for test production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build/allinone + '[' allinone '!=' _build ']' + cd .. + mkdir -p production + cd production + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + [[ true = true ]] + [[ true = true ]] + render_install_jaeger production-ui production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + '[' true = true ']' + INSECURE=true + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-forbbiden-access.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-curl.yaml + INSECURE=true + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./05-check-disabled-security.yaml + ASSERT_PRESENT=false + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./06-check-NO-gaID.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./08-check-gaID.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running ui E2E tests' Running ui E2E tests + cd tests/e2e/ui/_build + set +e + KUBECONFIG=/tmp/kubeconfig-998185858 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 3 tests === RUN kuttl/harness === RUN kuttl/harness/allinone === PAUSE kuttl/harness/allinone === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/production === PAUSE kuttl/harness/production === CONT kuttl/harness/allinone logger.go:42: 20:48:26 | allinone | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:48:26 | allinone | Creating namespace: kuttl-test-right-shepherd logger.go:42: 20:48:26 | allinone/0-install | starting test step 0-install logger.go:42: 20:48:26 | allinone/0-install | Jaeger:kuttl-test-right-shepherd/all-in-one-ui created logger.go:42: 20:48:30 | allinone/0-install | test step completed 0-install logger.go:42: 20:48:30 | allinone/1-curl | starting test step 1-curl logger.go:42: 20:48:30 | allinone/1-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 20:48:30 | allinone/1-curl | Checking the Ingress host value was populated logger.go:42: 20:48:30 | allinone/1-curl | Try number 0 logger.go:42: 20:48:30 | allinone/1-curl | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 20:48:30 | allinone/1-curl | template was: logger.go:42: 20:48:30 | allinone/1-curl | {.items[0].status.ingress[0].host} logger.go:42: 20:48:30 | allinone/1-curl | object given to jsonpath engine was: logger.go:42: 20:48:30 | allinone/1-curl | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 20:48:30 | allinone/1-curl | logger.go:42: 20:48:30 | allinone/1-curl | logger.go:42: 20:48:40 | allinone/1-curl | Try number 1 logger.go:42: 20:48:40 | allinone/1-curl | Hostname is all-in-one-ui-kuttl-test-right-shepherd.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com logger.go:42: 20:48:40 | allinone/1-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE all-in-one-ui] logger.go:42: 20:48:40 | allinone/1-curl | Checking an expected HTTP response logger.go:42: 20:48:40 | allinone/1-curl | Running in OpenShift logger.go:42: 20:48:40 | allinone/1-curl | User not provided. Getting the token... logger.go:42: 20:48:41 | allinone/1-curl | Warning: resource jaegers/all-in-one-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:48:48 | allinone/1-curl | Try number 1/30 the https://all-in-one-ui-kuttl-test-right-shepherd.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:48:48 | allinone/1-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 20:48:48 | allinone/1-curl | Try number 2/30 the https://all-in-one-ui-kuttl-test-right-shepherd.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:48:48 | allinone/1-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 20:48:58 | allinone/1-curl | Try number 3/30 the https://all-in-one-ui-kuttl-test-right-shepherd.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:48:58 | allinone/1-curl | curl response asserted properly logger.go:42: 20:48:58 | allinone/1-curl | test step completed 1-curl logger.go:42: 20:48:58 | allinone/2-delete | starting test step 2-delete logger.go:42: 20:48:58 | allinone/2-delete | Jaeger:kuttl-test-right-shepherd/all-in-one-ui created logger.go:42: 20:48:58 | allinone/2-delete | test step completed 2-delete logger.go:42: 20:48:58 | allinone/3-install | starting test step 3-install logger.go:42: 20:48:58 | allinone/3-install | Jaeger:kuttl-test-right-shepherd/all-in-one-ui updated logger.go:42: 20:48:58 | allinone/3-install | test step completed 3-install logger.go:42: 20:48:58 | allinone/4-test-ui-config | starting test step 4-test-ui-config logger.go:42: 20:48:58 | allinone/4-test-ui-config | running command: [./ensure-ingress-host.sh] logger.go:42: 20:48:58 | allinone/4-test-ui-config | Checking the Ingress host value was populated logger.go:42: 20:48:58 | allinone/4-test-ui-config | Try number 0 logger.go:42: 20:48:58 | allinone/4-test-ui-config | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 20:48:58 | allinone/4-test-ui-config | template was: logger.go:42: 20:48:58 | allinone/4-test-ui-config | {.items[0].status.ingress[0].host} logger.go:42: 20:48:58 | allinone/4-test-ui-config | object given to jsonpath engine was: logger.go:42: 20:48:58 | allinone/4-test-ui-config | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 20:48:58 | allinone/4-test-ui-config | logger.go:42: 20:48:58 | allinone/4-test-ui-config | logger.go:42: 20:49:08 | allinone/4-test-ui-config | Try number 1 logger.go:42: 20:49:09 | allinone/4-test-ui-config | Hostname is all-in-one-ui-kuttl-test-right-shepherd.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com logger.go:42: 20:49:09 | allinone/4-test-ui-config | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 20:49:09 | allinone/4-test-ui-config | time="2025-02-17T20:49:09Z" level=info msg="Querying https://all-in-one-ui-kuttl-test-right-shepherd.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search..." logger.go:42: 20:49:09 | allinone/4-test-ui-config | time="2025-02-17T20:49:09Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 20:49:09 | allinone/4-test-ui-config | time="2025-02-17T20:49:09Z" level=info msg="Polling to https://all-in-one-ui-kuttl-test-right-shepherd.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search" logger.go:42: 20:49:09 | allinone/4-test-ui-config | time="2025-02-17T20:49:09Z" level=info msg="Doing request number 0" logger.go:42: 20:49:09 | allinone/4-test-ui-config | time="2025-02-17T20:49:09Z" level=warning msg="Status code: 503" logger.go:42: 20:49:17 | allinone/4-test-ui-config | time="2025-02-17T20:49:17Z" level=info msg="Doing request number 1" logger.go:42: 20:49:17 | allinone/4-test-ui-config | time="2025-02-17T20:49:17Z" level=info msg="Content found and asserted!" logger.go:42: 20:49:17 | allinone/4-test-ui-config | time="2025-02-17T20:49:17Z" level=info msg="Success!" logger.go:42: 20:49:17 | allinone/4-test-ui-config | test step completed 4-test-ui-config logger.go:42: 20:49:17 | allinone | allinone events from ns kuttl-test-right-shepherd: logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:30 +0000 UTC Normal Pod all-in-one-ui-6d6c9f49f9-z6cjm Binding Scheduled Successfully assigned kuttl-test-right-shepherd/all-in-one-ui-6d6c9f49f9-z6cjm to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:30 +0000 UTC Normal Pod all-in-one-ui-6d6c9f49f9-z6cjm AddedInterface Add eth0 [10.128.2.71/23] from ovn-kubernetes multus logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:30 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-6d6c9f49f9 SuccessfulCreate Created pod: all-in-one-ui-6d6c9f49f9-z6cjm replicaset-controller logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:30 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-6d6c9f49f9 to 1 deployment-controller logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:31 +0000 UTC Normal Pod all-in-one-ui-6d6c9f49f9-z6cjm.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:31 +0000 UTC Normal Pod all-in-one-ui-6d6c9f49f9-z6cjm.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:31 +0000 UTC Normal Pod all-in-one-ui-6d6c9f49f9-z6cjm.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:31 +0000 UTC Normal Pod all-in-one-ui-6d6c9f49f9-z6cjm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:31 +0000 UTC Normal Pod all-in-one-ui-6d6c9f49f9-z6cjm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:31 +0000 UTC Normal Pod all-in-one-ui-6d6c9f49f9-z6cjm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:43 +0000 UTC Normal Pod all-in-one-ui-6d6c9f49f9-z6cjm.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:43 +0000 UTC Normal Pod all-in-one-ui-6d6c9f49f9-z6cjm.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:43 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-6d6c9f49f9 SuccessfulDelete Deleted pod: all-in-one-ui-6d6c9f49f9-z6cjm replicaset-controller logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:43 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-6d6c9f49f9 to 0 from 1 deployment-controller logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:44 +0000 UTC Normal Pod all-in-one-ui-79759d4f4-fvl4f Binding Scheduled Successfully assigned kuttl-test-right-shepherd/all-in-one-ui-79759d4f4-fvl4f to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:44 +0000 UTC Normal Pod all-in-one-ui-79759d4f4-fvl4f AddedInterface Add eth0 [10.128.2.72/23] from ovn-kubernetes multus logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:44 +0000 UTC Normal Pod all-in-one-ui-79759d4f4-fvl4f.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:44 +0000 UTC Normal Pod all-in-one-ui-79759d4f4-fvl4f.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:44 +0000 UTC Normal Pod all-in-one-ui-79759d4f4-fvl4f.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:44 +0000 UTC Normal Pod all-in-one-ui-79759d4f4-fvl4f.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:44 +0000 UTC Normal Pod all-in-one-ui-79759d4f4-fvl4f.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:44 +0000 UTC Normal Pod all-in-one-ui-79759d4f4-fvl4f.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:44 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-79759d4f4 SuccessfulCreate Created pod: all-in-one-ui-79759d4f4-fvl4f replicaset-controller logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:44 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-79759d4f4 to 1 deployment-controller logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:58 +0000 UTC Normal Pod all-in-one-ui-79759d4f4-fvl4f.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:48:58 +0000 UTC Normal Pod all-in-one-ui-79759d4f4-fvl4f.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:49:05 +0000 UTC Normal Pod all-in-one-ui-68b74c64cc-6xnm7 Binding Scheduled Successfully assigned kuttl-test-right-shepherd/all-in-one-ui-68b74c64cc-6xnm7 to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:49:17 | allinone | 2025-02-17 20:49:05 +0000 UTC Normal Pod all-in-one-ui-68b74c64cc-6xnm7 AddedInterface Add eth0 [10.129.2.72/23] from ovn-kubernetes multus logger.go:42: 20:49:17 | allinone | 2025-02-17 20:49:05 +0000 UTC Normal Pod all-in-one-ui-68b74c64cc-6xnm7.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:49:05 +0000 UTC Normal Pod all-in-one-ui-68b74c64cc-6xnm7.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:49:05 +0000 UTC Normal Pod all-in-one-ui-68b74c64cc-6xnm7.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:49:17 | allinone | 2025-02-17 20:49:05 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-68b74c64cc SuccessfulCreate Created pod: all-in-one-ui-68b74c64cc-6xnm7 replicaset-controller logger.go:42: 20:49:17 | allinone | 2025-02-17 20:49:05 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-68b74c64cc to 1 deployment-controller logger.go:42: 20:49:17 | allinone | Deleting namespace: kuttl-test-right-shepherd === CONT kuttl/harness/production logger.go:42: 20:49:24 | production | Ignoring add-tracking-id.yaml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:49:24 | production | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:49:24 | production | Creating namespace: kuttl-test-heroic-raven logger.go:42: 20:49:24 | production/1-install | starting test step 1-install logger.go:42: 20:49:24 | production/1-install | Jaeger:kuttl-test-heroic-raven/production-ui created logger.go:42: 20:50:01 | production/1-install | test step completed 1-install logger.go:42: 20:50:01 | production/2-check-forbbiden-access | starting test step 2-check-forbbiden-access logger.go:42: 20:50:01 | production/2-check-forbbiden-access | running command: [./ensure-ingress-host.sh] logger.go:42: 20:50:01 | production/2-check-forbbiden-access | Checking the Ingress host value was populated logger.go:42: 20:50:01 | production/2-check-forbbiden-access | Try number 0 logger.go:42: 20:50:01 | production/2-check-forbbiden-access | Hostname is production-ui-kuttl-test-heroic-raven.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com logger.go:42: 20:50:01 | production/2-check-forbbiden-access | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE production-ui] logger.go:42: 20:50:01 | production/2-check-forbbiden-access | Checking an expected HTTP response logger.go:42: 20:50:01 | production/2-check-forbbiden-access | Running in OpenShift logger.go:42: 20:50:01 | production/2-check-forbbiden-access | Not using any secret logger.go:42: 20:50:01 | production/2-check-forbbiden-access | Try number 1/30 the https://production-ui-kuttl-test-heroic-raven.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:50:01 | production/2-check-forbbiden-access | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 20:50:01 | production/2-check-forbbiden-access | Try number 2/30 the https://production-ui-kuttl-test-heroic-raven.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:50:01 | production/2-check-forbbiden-access | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 20:50:11 | production/2-check-forbbiden-access | Try number 3/30 the https://production-ui-kuttl-test-heroic-raven.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:50:11 | production/2-check-forbbiden-access | curl response asserted properly logger.go:42: 20:50:11 | production/2-check-forbbiden-access | test step completed 2-check-forbbiden-access logger.go:42: 20:50:11 | production/3-curl | starting test step 3-curl logger.go:42: 20:50:11 | production/3-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 20:50:11 | production/3-curl | Checking the Ingress host value was populated logger.go:42: 20:50:11 | production/3-curl | Try number 0 logger.go:42: 20:50:11 | production/3-curl | Hostname is production-ui-kuttl-test-heroic-raven.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com logger.go:42: 20:50:11 | production/3-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 20:50:12 | production/3-curl | Checking an expected HTTP response logger.go:42: 20:50:12 | production/3-curl | Running in OpenShift logger.go:42: 20:50:12 | production/3-curl | User not provided. Getting the token... logger.go:42: 20:50:13 | production/3-curl | Warning: resource jaegers/production-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:50:19 | production/3-curl | Try number 1/30 the https://production-ui-kuttl-test-heroic-raven.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:50:19 | production/3-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 20:50:19 | production/3-curl | Try number 2/30 the https://production-ui-kuttl-test-heroic-raven.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:50:19 | production/3-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 20:50:29 | production/3-curl | Try number 3/30 the https://production-ui-kuttl-test-heroic-raven.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:50:29 | production/3-curl | curl response asserted properly logger.go:42: 20:50:29 | production/3-curl | test step completed 3-curl logger.go:42: 20:50:29 | production/4-install | starting test step 4-install logger.go:42: 20:50:29 | production/4-install | Jaeger:kuttl-test-heroic-raven/production-ui updated logger.go:42: 20:50:29 | production/4-install | test step completed 4-install logger.go:42: 20:50:29 | production/5-check-disabled-security | starting test step 5-check-disabled-security logger.go:42: 20:50:29 | production/5-check-disabled-security | running command: [./ensure-ingress-host.sh] logger.go:42: 20:50:29 | production/5-check-disabled-security | Checking the Ingress host value was populated logger.go:42: 20:50:29 | production/5-check-disabled-security | Try number 0 logger.go:42: 20:50:30 | production/5-check-disabled-security | Hostname is production-ui-kuttl-test-heroic-raven.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com logger.go:42: 20:50:30 | production/5-check-disabled-security | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 20:50:30 | production/5-check-disabled-security | Checking an expected HTTP response logger.go:42: 20:50:30 | production/5-check-disabled-security | Running in OpenShift logger.go:42: 20:50:30 | production/5-check-disabled-security | Not using any secret logger.go:42: 20:50:30 | production/5-check-disabled-security | Try number 1/30 the https://production-ui-kuttl-test-heroic-raven.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:50:30 | production/5-check-disabled-security | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 20:50:30 | production/5-check-disabled-security | Try number 2/30 the https://production-ui-kuttl-test-heroic-raven.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:50:30 | production/5-check-disabled-security | HTTP response is 403. 200 expected. Waiting 10 s logger.go:42: 20:50:40 | production/5-check-disabled-security | Try number 3/30 the https://production-ui-kuttl-test-heroic-raven.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 20:50:40 | production/5-check-disabled-security | curl response asserted properly logger.go:42: 20:50:40 | production/5-check-disabled-security | test step completed 5-check-disabled-security logger.go:42: 20:50:40 | production/6-check-NO-gaID | starting test step 6-check-NO-gaID logger.go:42: 20:50:40 | production/6-check-NO-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 20:50:40 | production/6-check-NO-gaID | Checking the Ingress host value was populated logger.go:42: 20:50:40 | production/6-check-NO-gaID | Try number 0 logger.go:42: 20:50:40 | production/6-check-NO-gaID | Hostname is production-ui-kuttl-test-heroic-raven.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com logger.go:42: 20:50:40 | production/6-check-NO-gaID | running command: [sh -c ASSERT_PRESENT=false EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 20:50:40 | production/6-check-NO-gaID | time="2025-02-17T20:50:40Z" level=info msg="Querying https://production-ui-kuttl-test-heroic-raven.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search..." logger.go:42: 20:50:40 | production/6-check-NO-gaID | time="2025-02-17T20:50:40Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 20:50:40 | production/6-check-NO-gaID | time="2025-02-17T20:50:40Z" level=info msg="Polling to https://production-ui-kuttl-test-heroic-raven.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search" logger.go:42: 20:50:40 | production/6-check-NO-gaID | time="2025-02-17T20:50:40Z" level=info msg="Doing request number 0" logger.go:42: 20:50:40 | production/6-check-NO-gaID | time="2025-02-17T20:50:40Z" level=info msg="Content not found and asserted it was not found!" logger.go:42: 20:50:40 | production/6-check-NO-gaID | time="2025-02-17T20:50:40Z" level=info msg="Success!" logger.go:42: 20:50:40 | production/6-check-NO-gaID | test step completed 6-check-NO-gaID logger.go:42: 20:50:40 | production/7-add-tracking-id | starting test step 7-add-tracking-id logger.go:42: 20:50:40 | production/7-add-tracking-id | running command: [sh -c kubectl apply -f add-tracking-id.yaml -n $NAMESPACE] logger.go:42: 20:50:41 | production/7-add-tracking-id | jaeger.jaegertracing.io/production-ui configured logger.go:42: 20:50:41 | production/7-add-tracking-id | test step completed 7-add-tracking-id logger.go:42: 20:50:41 | production/8-check-gaID | starting test step 8-check-gaID logger.go:42: 20:50:41 | production/8-check-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 20:50:41 | production/8-check-gaID | Checking the Ingress host value was populated logger.go:42: 20:50:41 | production/8-check-gaID | Try number 0 logger.go:42: 20:50:41 | production/8-check-gaID | Hostname is production-ui-kuttl-test-heroic-raven.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com logger.go:42: 20:50:41 | production/8-check-gaID | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 20:50:41 | production/8-check-gaID | time="2025-02-17T20:50:41Z" level=info msg="Querying https://production-ui-kuttl-test-heroic-raven.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search..." logger.go:42: 20:50:41 | production/8-check-gaID | time="2025-02-17T20:50:41Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 20:50:41 | production/8-check-gaID | time="2025-02-17T20:50:41Z" level=info msg="Polling to https://production-ui-kuttl-test-heroic-raven.apps.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com/search" logger.go:42: 20:50:41 | production/8-check-gaID | time="2025-02-17T20:50:41Z" level=info msg="Doing request number 0" logger.go:42: 20:50:41 | production/8-check-gaID | time="2025-02-17T20:50:41Z" level=warning msg="Found: false . Assert: true" logger.go:42: 20:50:41 | production/8-check-gaID | time="2025-02-17T20:50:41Z" level=warning msg="The condition of the test function was not accomplished" logger.go:42: 20:50:49 | production/8-check-gaID | time="2025-02-17T20:50:49Z" level=info msg="Doing request number 1" logger.go:42: 20:51:09 | production/8-check-gaID | time="2025-02-17T20:51:09Z" level=info msg="Content found and asserted!" logger.go:42: 20:51:09 | production/8-check-gaID | time="2025-02-17T20:51:09Z" level=info msg="Success!" logger.go:42: 20:51:09 | production/8-check-gaID | test step completed 8-check-gaID logger.go:42: 20:51:09 | production | production events from ns kuttl-test-heroic-raven: logger.go:42: 20:51:09 | production | 2025-02-17 20:49:31 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestheroicravenproductionui-1-567945699d SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestheroicravenproductionui-1-56794ffj6h replicaset-controller logger.go:42: 20:51:09 | production | 2025-02-17 20:49:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestheroicravenproductionui-1-56794ffj6h Binding Scheduled Successfully assigned kuttl-test-heroic-raven/elasticsearch-cdm-kuttltestheroicravenproductionui-1-56794ffj6h to ip-10-0-89-201.us-east-2.compute.internal default-scheduler logger.go:42: 20:51:09 | production | 2025-02-17 20:49:31 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestheroicravenproductionui-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestheroicravenproductionui-1-567945699d to 1 deployment-controller logger.go:42: 20:51:09 | production | 2025-02-17 20:49:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestheroicravenproductionui-1-56794ffj6h AddedInterface Add eth0 [10.129.2.73/23] from ovn-kubernetes multus logger.go:42: 20:51:09 | production | 2025-02-17 20:49:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestheroicravenproductionui-1-56794ffj6h.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:5117d0aee260f74d6ef77c7c71ed073b1ad45b456f58894539b5697bec878310" already present on machine kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestheroicravenproductionui-1-56794ffj6h.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestheroicravenproductionui-1-56794ffj6h.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestheroicravenproductionui-1-56794ffj6h.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d63d612cdca72f64c45960a9b30bcfd0c746daabc17407df92c62694d5bc4bb6" already present on machine kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestheroicravenproductionui-1-56794ffj6h.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestheroicravenproductionui-1-56794ffj6h.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:42 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestheroicravenproductionui-1-56794ffj6h.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:47 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestheroicravenproductionui-1-56794ffj6h.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:58 +0000 UTC Normal Pod production-ui-collector-fb568bb88-mfvnb Binding Scheduled Successfully assigned kuttl-test-heroic-raven/production-ui-collector-fb568bb88-mfvnb to ip-10-0-44-238.us-east-2.compute.internal default-scheduler logger.go:42: 20:51:09 | production | 2025-02-17 20:49:58 +0000 UTC Normal Pod production-ui-collector-fb568bb88-mfvnb AddedInterface Add eth0 [10.131.0.75/23] from ovn-kubernetes multus logger.go:42: 20:51:09 | production | 2025-02-17 20:49:58 +0000 UTC Normal Pod production-ui-collector-fb568bb88-mfvnb.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:58 +0000 UTC Normal ReplicaSet.apps production-ui-collector-fb568bb88 SuccessfulCreate Created pod: production-ui-collector-fb568bb88-mfvnb replicaset-controller logger.go:42: 20:51:09 | production | 2025-02-17 20:49:58 +0000 UTC Normal Deployment.apps production-ui-collector ScalingReplicaSet Scaled up replica set production-ui-collector-fb568bb88 to 1 deployment-controller logger.go:42: 20:51:09 | production | 2025-02-17 20:49:58 +0000 UTC Normal Pod production-ui-query-84df8f476c-wph6h Binding Scheduled Successfully assigned kuttl-test-heroic-raven/production-ui-query-84df8f476c-wph6h to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:51:09 | production | 2025-02-17 20:49:58 +0000 UTC Normal Pod production-ui-query-84df8f476c-wph6h AddedInterface Add eth0 [10.128.2.73/23] from ovn-kubernetes multus logger.go:42: 20:51:09 | production | 2025-02-17 20:49:58 +0000 UTC Normal Pod production-ui-query-84df8f476c-wph6h.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:58 +0000 UTC Normal ReplicaSet.apps production-ui-query-84df8f476c SuccessfulCreate Created pod: production-ui-query-84df8f476c-wph6h replicaset-controller logger.go:42: 20:51:09 | production | 2025-02-17 20:49:58 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-84df8f476c to 1 deployment-controller logger.go:42: 20:51:09 | production | 2025-02-17 20:49:59 +0000 UTC Normal Pod production-ui-collector-fb568bb88-mfvnb.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:59 +0000 UTC Normal Pod production-ui-collector-fb568bb88-mfvnb.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:59 +0000 UTC Normal Pod production-ui-query-84df8f476c-wph6h.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:59 +0000 UTC Normal Pod production-ui-query-84df8f476c-wph6h.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:59 +0000 UTC Normal Pod production-ui-query-84df8f476c-wph6h.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:59 +0000 UTC Normal Pod production-ui-query-84df8f476c-wph6h.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:59 +0000 UTC Normal Pod production-ui-query-84df8f476c-wph6h.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:59 +0000 UTC Normal Pod production-ui-query-84df8f476c-wph6h.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:59 +0000 UTC Normal Pod production-ui-query-84df8f476c-wph6h.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:49:59 +0000 UTC Normal Pod production-ui-query-84df8f476c-wph6h.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:13 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:51:09 | production | 2025-02-17 20:50:13 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:51:09 | production | 2025-02-17 20:50:13 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:51:09 | production | 2025-02-17 20:50:14 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-84df8f476c to 0 from 1 deployment-controller logger.go:42: 20:51:09 | production | 2025-02-17 20:50:15 +0000 UTC Normal Pod production-ui-query-579878bd84-f9jq9 Binding Scheduled Successfully assigned kuttl-test-heroic-raven/production-ui-query-579878bd84-f9jq9 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:51:09 | production | 2025-02-17 20:50:15 +0000 UTC Normal ReplicaSet.apps production-ui-query-579878bd84 SuccessfulCreate Created pod: production-ui-query-579878bd84-f9jq9 replicaset-controller logger.go:42: 20:51:09 | production | 2025-02-17 20:50:15 +0000 UTC Normal Pod production-ui-query-84df8f476c-wph6h.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:15 +0000 UTC Normal Pod production-ui-query-84df8f476c-wph6h.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:15 +0000 UTC Normal Pod production-ui-query-84df8f476c-wph6h.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:15 +0000 UTC Normal ReplicaSet.apps production-ui-query-84df8f476c SuccessfulDelete Deleted pod: production-ui-query-84df8f476c-wph6h replicaset-controller logger.go:42: 20:51:09 | production | 2025-02-17 20:50:15 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-579878bd84 to 1 deployment-controller logger.go:42: 20:51:09 | production | 2025-02-17 20:50:16 +0000 UTC Normal Pod production-ui-query-579878bd84-f9jq9 AddedInterface Add eth0 [10.128.2.74/23] from ovn-kubernetes multus logger.go:42: 20:51:09 | production | 2025-02-17 20:50:16 +0000 UTC Normal Pod production-ui-query-579878bd84-f9jq9.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:16 +0000 UTC Normal Pod production-ui-query-579878bd84-f9jq9.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:16 +0000 UTC Normal Pod production-ui-query-579878bd84-f9jq9.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:16 +0000 UTC Normal Pod production-ui-query-579878bd84-f9jq9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:16 +0000 UTC Normal Pod production-ui-query-579878bd84-f9jq9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:16 +0000 UTC Normal Pod production-ui-query-579878bd84-f9jq9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:16 +0000 UTC Normal Pod production-ui-query-579878bd84-f9jq9.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:16 +0000 UTC Normal Pod production-ui-query-579878bd84-f9jq9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:16 +0000 UTC Normal Pod production-ui-query-579878bd84-f9jq9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:31 +0000 UTC Normal Pod production-ui-query-579878bd84-f9jq9.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:31 +0000 UTC Normal Pod production-ui-query-579878bd84-f9jq9.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:31 +0000 UTC Normal Pod production-ui-query-579878bd84-f9jq9.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:31 +0000 UTC Normal ReplicaSet.apps production-ui-query-579878bd84 SuccessfulDelete Deleted pod: production-ui-query-579878bd84-f9jq9 replicaset-controller logger.go:42: 20:51:09 | production | 2025-02-17 20:50:31 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-579878bd84 to 0 from 1 deployment-controller logger.go:42: 20:51:09 | production | 2025-02-17 20:50:32 +0000 UTC Normal Pod production-ui-query-78fd6476c4-nfzp8 Binding Scheduled Successfully assigned kuttl-test-heroic-raven/production-ui-query-78fd6476c4-nfzp8 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:51:09 | production | 2025-02-17 20:50:32 +0000 UTC Normal ReplicaSet.apps production-ui-query-78fd6476c4 SuccessfulCreate Created pod: production-ui-query-78fd6476c4-nfzp8 replicaset-controller logger.go:42: 20:51:09 | production | 2025-02-17 20:50:32 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-78fd6476c4 to 1 deployment-controller logger.go:42: 20:51:09 | production | 2025-02-17 20:50:33 +0000 UTC Normal Pod production-ui-query-78fd6476c4-nfzp8 AddedInterface Add eth0 [10.128.2.75/23] from ovn-kubernetes multus logger.go:42: 20:51:09 | production | 2025-02-17 20:50:33 +0000 UTC Normal Pod production-ui-query-78fd6476c4-nfzp8.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:33 +0000 UTC Normal Pod production-ui-query-78fd6476c4-nfzp8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:33 +0000 UTC Normal Pod production-ui-query-78fd6476c4-nfzp8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:33 +0000 UTC Normal Pod production-ui-query-78fd6476c4-nfzp8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:33 +0000 UTC Normal Pod production-ui-query-78fd6476c4-nfzp8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:33 +0000 UTC Normal Pod production-ui-query-78fd6476c4-nfzp8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:42 +0000 UTC Normal Pod production-ui-query-78fd6476c4-nfzp8.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:42 +0000 UTC Normal Pod production-ui-query-78fd6476c4-nfzp8.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:42 +0000 UTC Normal ReplicaSet.apps production-ui-query-78fd6476c4 SuccessfulDelete Deleted pod: production-ui-query-78fd6476c4-nfzp8 replicaset-controller logger.go:42: 20:51:09 | production | 2025-02-17 20:50:42 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-78fd6476c4 to 0 from 1 deployment-controller logger.go:42: 20:51:09 | production | 2025-02-17 20:50:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod production-ui-collector-fb568bb88-mfvnb horizontal-pod-autoscaler logger.go:42: 20:51:09 | production | 2025-02-17 20:50:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod production-ui-collector-fb568bb88-mfvnb horizontal-pod-autoscaler logger.go:42: 20:51:09 | production | 2025-02-17 20:50:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod production-ui-collector-fb568bb88-mfvnb horizontal-pod-autoscaler logger.go:42: 20:51:09 | production | 2025-02-17 20:50:43 +0000 UTC Warning Pod production-ui-query-78fd6476c4-nfzp8.spec.containers{jaeger-agent} Unhealthy Readiness probe failed: Get "http://10.128.2.75:14271/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:43 +0000 UTC Warning Pod production-ui-query-78fd6476c4-nfzp8.spec.containers{jaeger-query} Unhealthy Readiness probe failed: Get "http://10.128.2.75:16687/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:43 +0000 UTC Normal Pod production-ui-query-c6d976cc5-jmf89 Binding Scheduled Successfully assigned kuttl-test-heroic-raven/production-ui-query-c6d976cc5-jmf89 to ip-10-0-113-7.us-east-2.compute.internal default-scheduler logger.go:42: 20:51:09 | production | 2025-02-17 20:50:43 +0000 UTC Normal ReplicaSet.apps production-ui-query-c6d976cc5 SuccessfulCreate Created pod: production-ui-query-c6d976cc5-jmf89 replicaset-controller logger.go:42: 20:51:09 | production | 2025-02-17 20:50:43 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-c6d976cc5 to 1 deployment-controller logger.go:42: 20:51:09 | production | 2025-02-17 20:50:44 +0000 UTC Normal Pod production-ui-query-c6d976cc5-jmf89 AddedInterface Add eth0 [10.128.2.76/23] from ovn-kubernetes multus logger.go:42: 20:51:09 | production | 2025-02-17 20:50:44 +0000 UTC Normal Pod production-ui-query-c6d976cc5-jmf89.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:44 +0000 UTC Normal Pod production-ui-query-c6d976cc5-jmf89.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:44 +0000 UTC Normal Pod production-ui-query-c6d976cc5-jmf89.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:44 +0000 UTC Normal Pod production-ui-query-c6d976cc5-jmf89.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:44 +0000 UTC Normal Pod production-ui-query-c6d976cc5-jmf89.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:51:09 | production | 2025-02-17 20:50:44 +0000 UTC Normal Pod production-ui-query-c6d976cc5-jmf89.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:51:09 | production | Deleting namespace: kuttl-test-heroic-raven === CONT kuttl/harness/artifacts logger.go:42: 20:51:16 | artifacts | Creating namespace: kuttl-test-valid-foal logger.go:42: 20:51:16 | artifacts | artifacts events from ns kuttl-test-valid-foal: logger.go:42: 20:51:16 | artifacts | Deleting namespace: kuttl-test-valid-foal === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (176.86s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/allinone (57.99s) --- PASS: kuttl/harness/production (112.44s) --- PASS: kuttl/harness/artifacts (6.27s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name ui --report --output /logs/artifacts/ui.xml ./artifacts/kuttl-report.xml time="2025-02-17T20:51:23Z" level=debug msg="Setting a new name for the test suites" time="2025-02-17T20:51:23Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-02-17T20:51:23Z" level=debug msg="normalizing test case names" time="2025-02-17T20:51:23Z" level=debug msg="ui/allinone -> ui_allinone" time="2025-02-17T20:51:23Z" level=debug msg="ui/production -> ui_production" time="2025-02-17T20:51:23Z" level=debug msg="ui/artifacts -> ui_artifacts" +---------------+--------+ | NAME | RESULT | +---------------+--------+ | ui_allinone | passed | | ui_production | passed | | ui_artifacts | passed | +---------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh upgrade false true + '[' 3 -ne 3 ']' + test_suite_name=upgrade + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/upgrade.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-upgrade make[2]: Entering directory '/tmp/jaeger-tests' make docker JAEGER_VERSION=1.62.1 IMG="quay.io//jaeger-operator:next" make[3]: Entering directory '/tmp/jaeger-tests' [ ! -z "true" ] || docker build --build-arg=GOPROXY= --build-arg=VERSION="1.62.0" --build-arg=JAEGER_VERSION=1.62.1 --build-arg=TARGETARCH= --build-arg VERSION_DATE=2025-02-17T20:51:23Z --build-arg VERSION_PKG="github.com/jaegertracing/jaeger-operator/pkg/version" -t "quay.io//jaeger-operator:next" . make[3]: Leaving directory '/tmp/jaeger-tests' touch build-e2e-upgrade-image SKIP_ES_EXTERNAL=true IMG=quay.io//jaeger-operator:"1.62.0" JAEGER_OPERATOR_VERSION="1.62.0" JAEGER_VERSION="1.62.0" ./tests/e2e/upgrade/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-14-222249 True False 47m Cluster version is 4.18.0-0.nightly-2025-02-14-222249' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-14-222249 True False 47m Cluster version is 4.18.0-0.nightly-2025-02-14-222249' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/upgrade/render.sh ++ export SUITE_DIR=./tests/e2e/upgrade ++ SUITE_DIR=./tests/e2e/upgrade ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/upgrade ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + export JAEGER_NAME + '[' true = true ']' + skip_test upgrade 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade + warning 'upgrade: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade: Test not supported in OpenShift\e[0m' WAR: upgrade: Test not supported in OpenShift + '[' true = true ']' + skip_test upgrade-from-latest-release 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade-from-latest-release + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade-from-latest-release + warning 'upgrade-from-latest-release: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade-from-latest-release: Test not supported in OpenShift\e[0m' WAR: upgrade-from-latest-release: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running upgrade E2E tests' Running upgrade E2E tests + cd tests/e2e/upgrade/_build + set +e + KUBECONFIG=/tmp/kubeconfig-998185858 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-4n8v34gs-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 20:51:24 | artifacts | Creating namespace: kuttl-test-fine-elephant logger.go:42: 20:51:24 | artifacts | artifacts events from ns kuttl-test-fine-elephant: logger.go:42: 20:51:24 | artifacts | Deleting namespace: kuttl-test-fine-elephant === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.44s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.30s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name upgrade --report --output /logs/artifacts/upgrade.xml ./artifacts/kuttl-report.xml time="2025-02-17T20:51:31Z" level=debug msg="Setting a new name for the test suites" time="2025-02-17T20:51:31Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-02-17T20:51:31Z" level=debug msg="normalizing test case names" time="2025-02-17T20:51:31Z" level=debug msg="upgrade/artifacts -> upgrade_artifacts" +-------------------+--------+ | NAME | RESULT | +-------------------+--------+ | upgrade_artifacts | passed | +-------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests'